WorldWideScience

Sample records for monitor required software

  1. Modular Software Performance Monitoring

    CERN Document Server

    Kruse, D F

    2011-01-01

    CPU clock frequency is not likely to be increased significantly in the coming years, and data analysis speed can be improved by using more processors or buying new machines, only if one is willing to change the paradigm to a parallel one. Therefore, performance monitoring procedures and tools are needed to help programmers to optimize existing software running on current and future hardware. Low level information from hardware performance counters is vital to spot specific performance problems slowing program execution. HEP software is often huge and complex, and existing tools are unable to give results with the required granularity. We will report on the approach we have chose to solve this problem that involves decomposing the application into parts and monitoring each of them separately. Both counting and sampling methods are used to allow an analysis with the required custom granularity: from global level, up to the function level. A set of tools (based on perfmon2 – a software interface to hardware co...

  2. Automated Cryocooler Monitor and Control System Software

    Science.gov (United States)

    Britchcliffe, Michael J.; Conroy, Bruce L.; Anderson, Paul E.; Wilson, Ahmad

    2011-01-01

    This software is used in an automated cryogenic control system developed to monitor and control the operation of small-scale cryocoolers. The system was designed to automate the cryogenically cooled low-noise amplifier system described in "Automated Cryocooler Monitor and Control System" (NPO-47246), NASA Tech Briefs, Vol. 35, No. 5 (May 2011), page 7a. The software contains algorithms necessary to convert non-linear output voltages from the cryogenic diode-type thermometers and vacuum pressure and helium pressure sensors, to temperature and pressure units. The control function algorithms use the monitor data to control the cooler power, vacuum solenoid, vacuum pump, and electrical warm-up heaters. The control algorithms are based on a rule-based system that activates the required device based on the operating mode. The external interface is Web-based. It acts as a Web server, providing pages for monitor, control, and configuration. No client software from the external user is required.

  3. Software requirements elicitation to support internal monitoring of quality assurance system for higher education in Indonesia

    Science.gov (United States)

    Amalia, A.; Gunawan, D.; Hardi, S. M.; Rachmawati, D.

    2018-02-01

    The Internal Quality Assurance System (in Indonesian: SPMI (Sistem Penjaminan Mutu Internal) is a systemic activity of quality assurance of higher education in Indonesia. SPMI should be done by all higher education or universities in Indonesia based on the Regulation of the Minister of Research, Technology and Higher Education of the Republic of Indonesia Number 62 of 2016. Implementation of SPMI must refer to the principle of SPMI that is independent, standardize, accurate, well planned and sustainable, documented and systematic. To assist the SPMI cycle properly, universities need a supporting software to monitor all the activities of SPMI. But in reality, many universities are not optimal in building this SPMI monitoring system. One of the obstacles is the determination of system requirements in support of SPMI principles is difficult to achieve. In this paper, we observe the initial phase of the engineering requirements elicitation. Unlike other methods that collect system requirements from users and stakeholders, we find the system requirements of the SPMI principles from SPMI guideline book. The result of this paper can be used as a choice in determining SPMI software requirements. This paper can also be used by developers and users to understand the scenario of SPMI so that could overcome the problems of understanding between this two parties.

  4. Software requirements

    CERN Document Server

    Wiegers, Karl E

    2003-01-01

    Without formal, verifiable software requirements-and an effective system for managing them-the programs that developers think they've agreed to build often will not be the same products their customers are expecting. In SOFTWARE REQUIREMENTS, Second Edition, requirements engineering authority Karl Wiegers amplifies the best practices presented in his original award-winning text?now a mainstay for anyone participating in the software development process. In this book, you'll discover effective techniques for managing the requirements engineering process all the way through the development cy

  5. Mathis software for controlling BCAM-based monitoring and alignment systems

    CERN Document Server

    Klumb, Francis; Kautzmann, Guillaume; CERN. Geneva. ATS Department

    2016-01-01

    The MATHIS Software (Monitoring and Alignment Tracking for HIE-Isolde Software) aims at providing 3D positions of physical components of the HIE-Isolde superconducting modules, accurately and permanently measured by well-designed networks of BCAM devices (Brandeis Camera Angle Monitoring). Although it is originally intended for the HIE-Isolde project, its architecture and its use cases have been extended and optimized for more general setups. Most of the configuration data are stored either within XML-formatted files or within databases. The adaptation of MATHIS for different BCAM monitoring systems therefore does not require any further code rewriting. Moreover, the software is fully cross-platform and can either be run on the specific Linux machines driving the accelerator electronic devices, or be used on independent Windows workstations as a stand-alone software. In the first case, the software mainly relies on FESA (Front End Software Architecture) which is an object-oriented real-time framework that ens...

  6. A Framework for Evaluating the Software Product Quality of Pregnancy Monitoring Mobile Personal Health Records.

    Science.gov (United States)

    Idri, Ali; Bachiri, Mariam; Fernández-Alemán, José Luis

    2016-03-01

    Stakeholders' needs and expectations are identified by means of software quality requirements, which have an impact on software product quality. In this paper, we present a set of requirements for mobile personal health records (mPHRs) for pregnancy monitoring, which have been extracted from literature and existing mobile apps on the market. We also use the ISO/IEC 25030 standard to suggest the requirements that should be considered during the quality evaluation of these mPHRs. We then go on to design a checklist in which we contrast the mPHRs for pregnancy monitoring requirements with software product quality characteristics and sub-characteristics in order to calculate the impact of these requirements on software product quality, using the ISO/IEC 25010 software product quality standard. The results obtained show that the requirements related to the user's actions and the app's features have the most impact on the external sub-characteristics of the software product quality model. The only sub-characteristic affected by all the requirements is Appropriateness of Functional suitability. The characteristic Operability is affected by 95% of the requirements while the lowest degrees of impact were identified for Compatibility (15%) and Transferability (6%). Lastly, the degrees of the impact of the mPHRs for pregnancy monitoring requirements are discussed in order to provide appropriate recommendations for the developers and stakeholders of mPHRs for pregnancy monitoring.

  7. Functional requirements for gas characterization system computer software

    International Nuclear Information System (INIS)

    Tate, D.D.

    1996-01-01

    This document provides the Functional Requirements for the Computer Software operating the Gas Characterization System (GCS), which monitors the combustible gasses in the vapor space of selected tanks. Necessary computer functions are defined to support design, testing, operation, and change control. The GCS requires several individual computers to address the control and data acquisition functions of instruments and sensors. These computers are networked for communication, and must multi-task to accommodate operation in parallel

  8. Neutrosophy for software requirement prioritization

    Directory of Open Access Journals (Sweden)

    Ronald Barriga Dias

    2017-09-01

    Full Text Available Software engineers are involved in complex decisions that require multiples viewpoints. A specific case is the requirement prioritization process. This process is used to decide which software requirement to develop in certain release from a group of candidate requirements. Criteria involved in this process can involve indeterminacy. In this paper a software requirement prioritization model is develop based SVN numbers. Finally, an illustrative example is presented in order to show the proposed model.

  9. Treated effluent disposal system process control computer software requirements and specification

    International Nuclear Information System (INIS)

    Graf, F.A. Jr.

    1994-01-01

    The software requirements for the monitor and control system that will be associated with the effluent collection pipeline system known as the 200 Area Treated Effluent Disposal System is covered. The control logic for the two pump stations and specific requirements for the graphic displays are detailed

  10. UTM TCL2 Software Requirements

    Science.gov (United States)

    Smith, Irene S.; Rios, Joseph L.; McGuirk, Patrick O.; Mulfinger, Daniel G.; Venkatesan, Priya; Smith, David R.; Baskaran, Vijayakumar; Wang, Leo

    2017-01-01

    The Unmanned Aircraft Systems (UAS) Traffic Management (UTM) Technical Capability Level (TCL) 2 software implements the UTM TCL 2 software requirements described herein. These software requirements are linked to the higher level UTM TCL 2 System Requirements. Each successive TCL implements additional UTM functionality, enabling additional use cases. TCL 2 demonstrated how to enable expanded multiple operations by implementing automation for beyond visual line-of-sight, tracking operations, and operations flying over sparsely populated areas.

  11. Verification of operating software for cooperative monitoring applications

    International Nuclear Information System (INIS)

    Tolk, K.M.; Rembold, R.K.

    1997-01-01

    Monitoring agencies often use computer based equipment to control instruments and to collect data at sites that are being monitored under international safeguards or other cooperative monitoring agreements. In order for this data to be used as an independent verification of data supplied by the host at the facility, the software used must be trusted by the monitoring agency. The monitoring party must be sure that the software has not be altered to give results that could lead to erroneous conclusions about nuclear materials inventories or other operating conditions at the site. The host might also want to verify that the software being used is the software that has been previously inspected in order to be assured that only data that is allowed under the agreement is being collected. A description of a method to provide this verification using keyed has functions and how the proposed method overcomes possible vulnerabilities in methods currently in use such as loading the software from trusted disks is presented. The use of public key data authentication for this purpose is also discussed

  12. Software development of the KSTAR Tokamak Monitoring System

    International Nuclear Information System (INIS)

    Kim, K.H.; Lee, T.G.; Baek, S.; Lee, S.I.; Chu, Y.; Kim, Y.O.; Kim, J.S.; Park, M.K.; Oh, Y.K.

    2008-01-01

    The Korea Superconducting Tokamak Advanced Research (KSTAR) project, which is constructing a superconducting Tokamak, was launched in 1996. Much progress in instrumentation and control has been made since then and the construction phase will be finished in August 2007. The Tokamak Monitoring System (TMS) measures the temperatures of the superconducting magnets, bus-lines, and structures and hence monitors the superconducting conditions during the operation of the KSTAR Tokamak. The TMS also measures the strains and displacements on the structures in order to monitor the mechanical safety. There are around 400 temperature sensors, more than 240 strain gauges, 10 displacement gauges and 10 Hall sensors. The TMS utilizes Cernox sensors for low temperature measurement and each sensor has its own characteristic curve. In addition, the TMS needs to perform complex arithmetic operations to convert the measurements into temperatures for each Cernox sensor for this large number of monitoring channels. A special software development effort was required to reduce the temperature conversion time and multi-threading to achieve the higher performance needed to handle the large number of channels. We have developed the TMS with PXI hardware and with EPICS software. We will describe the details of the implementations in this paper

  13. CLOUDCLOUD : general-purpose instrument monitoring and data managing software

    Science.gov (United States)

    Dias, António; Amorim, António; Tomé, António

    2016-04-01

    monitored with parsing intervals as fast as milliseconds. This software (server+agents+interface+database) comes in easy and ready-to-use packages that can be installed in any operating system, including Android and iOS systems. This software is ideal for use in modular experiments or monitoring stations with large variability in instruments and measuring methods or in large collaborations, where data requires homogenization in order to be effectively transmitted to all involved parties. This work presents the software and provides performance comparison with previously used monitoring systems in the CLOUD experiment at CERN.

  14. Monitoring extensions for component-based distributed software

    NARCIS (Netherlands)

    Diakov, N.K.; Papir, Z.; van Sinderen, Marten J.; Quartel, Dick

    2000-01-01

    This paper defines a generic class of monitoring extensions to component-based distributed enterprise software. Introducing a monitoring extension to a legacy application system can be very costly. In this paper, we identify the minimum support for application monitoring within the generic

  15. Requirements Engineering for Software Integrity and Safety

    Science.gov (United States)

    Leveson, Nancy G.

    2002-01-01

    Requirements flaws are the most common cause of errors and software-related accidents in operational software. Most aerospace firms list requirements as one of their most important outstanding software development problems and all of the recent, NASA spacecraft losses related to software (including the highly publicized Mars Program failures) can be traced to requirements flaws. In light of these facts, it is surprising that relatively little research is devoted to requirements in contrast with other software engineering topics. The research proposed built on our previous work. including both criteria for determining whether a requirements specification is acceptably complete and a new approach to structuring system specifications called Intent Specifications. This grant was to fund basic research on how these ideas could be extended to leverage innovative approaches to the problems of (1) reducing the impact of changing requirements, (2) finding requirements specification flaws early through formal and informal analysis, and (3) avoiding common flaws entirely through appropriate requirements specification language design.

  16. Writing testable software requirements

    Energy Technology Data Exchange (ETDEWEB)

    Knirk, D. [Sandia National Labs., Albuquerque, NM (United States)

    1997-11-01

    This tutorial identifies common problems in analyzing requirements in the problem and constructing a written specification of what the software is to do. It deals with two main problem areas: identifying and describing problem requirements, and analyzing and describing behavior specifications.

  17. Software for airborne radiation monitoring system

    International Nuclear Information System (INIS)

    Sheinfeld, M.; Kadmon, Y.; Tirosh, D.; Elhanany, I.; Gabovitch, A.; Barak, D.

    1997-01-01

    The Airborne Radiation Monitoring System monitors radioactive contamination in the air or on the ground. The contamination source can be a radioactive plume or an area contaminated with radionuclides. This system is composed of two major parts: Airborne Unit carried by a helicopter, and Ground Station carried by a truck. The Airborne software is intended to be the core of a computerized airborne station. The software is written in C++ under MS-Windows with object-oriented methodology. It has been designed to be user-friendly: function keys and other accelerators are used for vital operations, a help file and help subjects are available, the Human-Machine-Interface is plain and obvious. (authors)

  18. Aligning Requirements-Driven Software Processes with IT Governance

    OpenAIRE

    Nguyen Huynh Anh, Vu; Kolp, Manuel; Heng, Samedi; Wautelet, Yves

    2017-01-01

    Requirements Engineering is closely intertwined with Information Technology (IT) Governance. Aligning IT Governance principles with Requirements-Driven Software Processes allows them to propose governance and management rules for software development to cope with stakeholders’ requirements and expectations. Typically, the goal of IT Governance in software engineering is to ensure that the results of a software organization business processes meet the strategic requirements of the organization...

  19. Requirements engineering: foundation for software quality

    NARCIS (Netherlands)

    Daneva, Maia; Pastor, Oscar

    2016-01-01

    Welcome to the proceedings of the 22nd edition of REFSQ: the International Working Conference on Requirements Engineering – Foundation for Software Quality! Requirements engineering (RE) has been recognized as a critical factor that impacts the quality of software, systems, and services. Since the

  20. The family of standard hydrogen monitoring system computer software design description: Revision 2

    International Nuclear Information System (INIS)

    Bender, R.M.

    1994-01-01

    In March 1990, 23 waste tanks at the Hanford Nuclear Reservation were identified as having the potential for the buildup of gas to a flammable or explosive level. As a result of the potential for hydrogen gas buildup, a project was initiated to design a standard hydrogen monitoring system (SHMS) for use at any waste tank to analyze gas samples for hydrogen content. Since it was originally deployed three years ago, two variations of the original system have been developed: the SHMS-B and SHMS-C. All three are currently in operation at the tank farms and will be discussed in this document. To avoid confusion in this document, when a feature is common to all three of the SHMS variants, it will be referred to as ''The family of SHMS.'' When it is specific to only one or two, they will be identified. The purpose of this computer software design document is to provide the following: the computer software requirements specification that documents the essential requirements of the computer software and its external interfaces; the computer software design description; the computer software user documentation for using and maintaining the computer software and any dedicated hardware; and the requirements for computer software design verification and validation

  1. An effective technique for the software requirements analysis of NPP safety-critical systems, based on software inspection, requirements traceability, and formal specification

    International Nuclear Information System (INIS)

    Koo, Seo Ryong; Seong, Poong Hyun; Yoo, Junbeom; Cha, Sung Deok; Yoo, Yeong Jae

    2005-01-01

    A thorough requirements analysis is indispensable for developing and implementing safety-critical software systems such as nuclear power plant (NPP) software systems because a single error in the requirements can generate serious software faults. However, it is very difficult to completely analyze system requirements. In this paper, an effective technique for the software requirements analysis is suggested. For requirements verification and validation (V and V) tasks, our technique uses software inspection, requirement traceability, and formal specification with structural decomposition. Software inspection and requirements traceability analysis are widely considered the most effective software V and V methods. Although formal methods are also considered an effective V and V activity, they are difficult to use properly in the nuclear fields as well as in other fields because of their mathematical nature. In this work, we propose an integrated environment (IE) approach for requirements, which is an integrated approach that enables easy inspection by combining requirement traceability and effective use of a formal method. The paper also introduces computer-aided tools for supporting IE approach for requirements. Called the nuclear software inspection support and requirements traceability (NuSISRT), the tool incorporates software inspection, requirement traceability, and formal specification capabilities. We designed the NuSISRT to partially automate software inspection and analysis of requirement traceability. In addition, for the formal specification and analysis, we used the formal requirements specification and analysis tool for nuclear engineering (NuSRS)

  2. Developing Software For Monitoring And Diagnosis

    Science.gov (United States)

    Edwards, S. J.; Caglayan, A. K.

    1993-01-01

    Expert-system software shell produces executable code. Report discusses beginning phase of research directed toward development of artificial intelligence for real-time monitoring of, and diagnosis of faults in, complicated systems of equipment. Motivated by need for onboard monitoring and diagnosis of electronic sensing and controlling systems of advanced aircraft. Also applicable to such equipment systems as refineries, factories, and powerplants.

  3. Software structure for tritium-in-air monitoring in classified locations

    International Nuclear Information System (INIS)

    Ionete, Eusebiu Ilarian; Benchea, Dumitru

    2009-01-01

    Full text: In the working areas of heavy water detritiation facilities, were hydrogen gas is generated, the risk of air-hydrogen explosive mixture production is present. This paper gives the description of a software architecture solution for a fixed area tritium-in-air monitoring system suitable to be used in such hazardous locations. Tritium-in-air monitoring system was designed as a distributed system containing a number of fixed tritium in air monitoring units, each of them being composed of an ionisation chamber flow-through type and a fix sampling unit with a number of sample lines and one additional blow line. For each unit, software and hardware architecture structure enable independent performance with a fail-safe concept, remote control operation and data storage in a DCS-DCU module. This software architecture secures the setting possibility of alarm levels for tritium concentration, graphical visualisation and acoustic alarm in the case of level overtaking. The afferent software architecture contains applications in connection with hardware architecture and with a hydrogen detection system: RU-1 software application for control sampling; RU-2 software application for measured values display; RU-3 software application for remote data display; RU-4 software application for SQL conversion server; RU-5 software application for OPC standard conversion. The interconnection of all hardware components, between each PLC and each display unit, between tritium-in-air monitoring units and PLC sampling units will be internally carried out using interfaces. The interconnection between PLCs and tritium-in-air monitoring DCU will be made using a communication network, for instance Ethernet Profibus or RS 485. (authors)

  4. IMAGE information monitoring and applied graphics software environment. Volume 2. Software description

    International Nuclear Information System (INIS)

    Hallam, J.W.; Ng, K.B.; Upham, G.L.

    1986-09-01

    The EPRI Information Monitoring and Applied Graphics Environment (IMAGE) system is designed for 'fast proto-typing' of advanced concepts for computer-aided plant operations tools. It is a flexible software system which can be used for rapidly creating, dynamically driving and evaluating advanced operator aid displays. The software is written to be both host computer and graphic device independent

  5. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    Batinkov, Atanas Ivanov

    2013-01-01

    The Data Quality Monitoring Software proved to be a central tool in the Compact Muon Solenoid experiment. Its flexibility allowed its integration in several environments: online, for real-time detector monitoring; offline, for the final, fine-grained data certification. The usage of the Data Quality Monitoring software in the different environments and its integration in the Compact Muon Solenoid reconstruction software framework and in all production workflows are presented. The main technical challenges and the adopted solutions to them will be also discussed with emphasis on functionality, long-term robustness and performance.

  6. The NLC Software Requirements Methodology

    Energy Technology Data Exchange (ETDEWEB)

    Shoaee, Hamid

    2002-08-20

    We describe the software requirements and development methodology developed for the NLC control system. Given the longevity of that project, and the likely geographical distribution of the collaborating engineers, the planned requirements management process is somewhat more formal than the norm in high energy physics projects. The short term goals of the requirements process are to accurately estimate costs, to decompose the problem, and to determine likely technologies. The long term goal is to enable a smooth transition from high level functional requirements to specific subsystem and component requirements for individual programmers, and to support distributed development. The methodology covers both ends of that life cycle. It covers both the analytical and documentary tools for software engineering, and project management support. This paper introduces the methodology, which is fully described in [1].

  7. Trend Monitoring System (TMS) graphics software

    Science.gov (United States)

    Brown, J. S.

    1979-01-01

    A prototype bus communications systems, which is being used to support the Trend Monitoring System (TMS) and to evaluate the bus concept is considered. A set of FORTRAN-callable graphics subroutines for the host MODCOMP comuter, and an approach to splitting graphics work between the host and the system's intelligent graphics terminals are described. The graphics software in the MODCOMP and the operating software package written for the graphics terminals are included.

  8. Health Monitor for Multitasking, Safety-Critical, Real-Time Software

    Science.gov (United States)

    Zoerner, Roger

    2011-01-01

    Health Manager can detect Bad Health prior to a failure occurring by periodically monitoring the application software by looking for code corruption errors, and sanity-checking each critical data value prior to use. A processor s memory can fail and corrupt the software, or the software can accidentally write to the wrong address and overwrite the executing software. This innovation will continuously calculate a checksum of the software load to detect corrupted code. This will allow a system to detect a failure before it happens. This innovation monitors each software task (thread) so that if any task reports "bad health," or does not report to the Health Manager, the system is declared bad. The Health Manager reports overall system health to the outside world by outputting a square wave signal. If the square wave stops, this indicates that system health is bad or hung and cannot report. Either way, "bad health" can be detected, whether caused by an error, corrupted data, or a hung processor. A separate Health Monitor Task is started and run periodically in a loop that starts and stops pending on a semaphore. Each monitored task registers with the Health Manager, which maintains a count for the task. The registering task must indicate if it will run more or less often than the Health Manager. If the task runs more often than the Health Manager, the monitored task calls a health function that increments the count and verifies it did not go over max-count. When the periodic Health Manager runs, it verifies that the count did not go over the max-count and zeroes it. If the task runs less often than the Health Manager, the periodic Health Manager will increment the count. The monitored task zeroes the count, and both the Health Manager and monitored task verify that the count did not go over the max-count.

  9. A Quantitative Study of Global Software Development Teams, Requirements, and Software Projects

    Science.gov (United States)

    Parker, Linda L.

    2016-01-01

    The study explored the relationship between global software development teams, effective software requirements, and stakeholders' perception of successful software development projects within the field of information technology management. It examined the critical relationship between Global Software Development (GSD) teams creating effective…

  10. Software attribute visualization for high integrity software

    Energy Technology Data Exchange (ETDEWEB)

    Pollock, G.M.

    1998-03-01

    This report documents a prototype tool developed to investigate the use of visualization and virtual reality technologies for improving software surety confidence. The tool is utilized within the execution phase of the software life cycle. It provides a capability to monitor an executing program against prespecified requirements constraints provided in a program written in the requirements specification language SAGE. The resulting Software Attribute Visual Analysis Tool (SAVAnT) also provides a technique to assess the completeness of a software specification.

  11. Runtime Performance Monitoring Tool for RTEMS System Software

    Science.gov (United States)

    Cho, B.; Kim, S.; Park, H.; Kim, H.; Choi, J.; Chae, D.; Lee, J.

    2007-08-01

    RTEMS is a commercial-grade real-time operating system that supports multi-processor computers. However, there are not many development tools for RTEMS. In this paper, we report new RTEMS-based runtime performance monitoring tool. We have implemented a light weight runtime monitoring task with an extension to the RTEMS APIs. Using our tool, software developers can verify various performance- related parameters during runtime. Our tool can be used during software development phase and in-orbit operation as well. Our implemented target agent is light weight and has small overhead using SpaceWire interface. Efforts to reduce overhead and to add other monitoring parameters are currently under research.

  12. Tank Monitoring and Document control System (TMACS) As Built Software Design Document

    International Nuclear Information System (INIS)

    GLASSCOCK, J.A.

    2000-01-01

    This document describes the software design for the Tank Monitor and Control System (TMACS). This document captures the existing as-built design of TMACS as of November 1999. It will be used as a reference document to the system maintainers who will be maintaining and modifying the TMACS functions as necessary. The heart of the TMACS system is the ''point-processing'' functionality where a sample value is received from the field sensors and the value is analyzed, logged, or alarmed as required. This Software Design Document focuses on the point-processing functions

  13. Requirements Engineering in Building Climate Science Software

    Science.gov (United States)

    Batcheller, Archer L.

    Software has an important role in supporting scientific work. This dissertation studies teams that build scientific software, focusing on the way that they determine what the software should do. These requirements engineering processes are investigated through three case studies of climate science software projects. The Earth System Modeling Framework assists modeling applications, the Earth System Grid distributes data via a web portal, and the NCAR (National Center for Atmospheric Research) Command Language is used to convert, analyze and visualize data. Document analysis, observation, and interviews were used to investigate the requirements-related work. The first research question is about how and why stakeholders engage in a project, and what they do for the project. Two key findings arise. First, user counts are a vital measure of project success, which makes adoption important and makes counting tricky and political. Second, despite the importance of quantities of users, a few particular "power users" develop a relationship with the software developers and play a special role in providing feedback to the software team and integrating the system into user practice. The second research question focuses on how project objectives are articulated and how they are put into practice. The team seeks to both build a software system according to product requirements but also to conduct their work according to process requirements such as user support. Support provides essential communication between users and developers that assists with refining and identifying requirements for the software. It also helps users to learn and apply the software to their real needs. User support is a vital activity for scientific software teams aspiring to create infrastructure. The third research question is about how change in scientific practice and knowledge leads to changes in the software, and vice versa. The "thickness" of a layer of software infrastructure impacts whether the

  14. Software system development of NPP plant DiD risk monitor. Basic design of software configuration

    International Nuclear Information System (INIS)

    Yoshikawa, Hidekazu; Nakagawa, Takashi

    2015-01-01

    A new risk monitor system is under development which can be applied not only to prevent severe accident in daily operation but also to serve as to mitigate the radiological hazard just after severe accident happens and long term management of post-severe accident consequences. The fundamental method for the new risk monitor system is first given on how to configure the Plant Defense in-Depth (DiD) Risk Monitor by object-oriented software system based on functional modeling approach. In this paper, software system for the plant DiD risk monitor is newly developed by object oriented method utilizing Unified Modeling Language (UML). Usage of the developed DiD risk monitor is also introduced by showing examples for LOCA case of AP1000. (author)

  15. FFI: A software tool for ecological monitoring

    Science.gov (United States)

    Duncan C. Lutes; Nathan C. Benson; MaryBeth Keifer; John F. Caratti; S. Austin Streetman

    2009-01-01

    A new monitoring tool called FFI (FEAT/FIREMON Integrated) has been developed to assist managers with collection, storage and analysis of ecological information. The tool was developed through the complementary integration of two fire effects monitoring systems commonly used in the United States: FIREMON and the Fire Ecology Assessment Tool. FFI provides software...

  16. Requirements: Towards an understanding on why software projects fail

    Science.gov (United States)

    Hussain, Azham; Mkpojiogu, Emmanuel O. C.

    2016-08-01

    Requirement engineering is at the foundation of every successful software project. There are many reasons for software project failures; however, poorly engineered requirements process contributes immensely to the reason why software projects fail. Software project failure is usually costly and risky and could also be life threatening. Projects that undermine requirements engineering suffer or are likely to suffer from failures, challenges and other attending risks. The cost of project failures and overruns when estimated is very huge. Furthermore, software project failures or overruns pose a challenge in today's competitive market environment. It affects the company's image, goodwill, and revenue drive and decreases the perceived satisfaction of customers and clients. In this paper, requirements engineering was discussed. Its role in software projects success was elaborated. The place of software requirements process in relation to software project failure was explored and examined. Also, project success and failure factors were also discussed with emphasis placed on requirements factors as they play a major role in software projects' challenges, successes and failures. The paper relied on secondary data and empirical statistics to explore and examine factors responsible for the successes, challenges and failures of software projects in large, medium and small scaled software companies.

  17. Monitoring Requirements Coverage Using Reconstructed Views : An Industrial Case Study

    NARCIS (Netherlands)

    Lormans, M.; Gross, H.; Van Deursen, A.; Van Solingen, R.; Stehouwer, A.

    2006-01-01

    Requirements views, such as coverage and status views, are an important asset for monitoring and managing software development. We have developed a method that automates the process for reconstructing these views, and built a tool, ReqAnalyst, to support this method. In this paper, we investigate to

  18. 77 FR 50726 - Software Requirement Specifications for Digital Computer Software and Complex Electronics Used in...

    Science.gov (United States)

    2012-08-22

    ... Computer Software and Complex Electronics Used in Safety Systems of Nuclear Power Plants AGENCY: Nuclear...-1209, ``Software Requirement Specifications for Digital Computer Software and Complex Electronics used... Electronics Engineers (ANSI/IEEE) Standard 830-1998, ``IEEE Recommended Practice for Software Requirements...

  19. The architecture of a reliable software monitoring system for embedded software systems

    International Nuclear Information System (INIS)

    Munson, J.; Krings, A.; Hiromoto, R.

    2006-01-01

    We develop the notion of a measurement-based methodology for embedded software systems to ensure properties of reliability, survivability and security, not only under benign faults but under malicious and hazardous conditions as well. The driving force is the need to develop a dynamic run-time monitoring system for use in these embedded mission critical systems. These systems must run reliably, must be secure and they must fail gracefully. That is, they must continue operating in the face of the departures from their nominal operating scenarios, the failure of one or more system components due to normal hardware and software faults, as well as malicious acts. To insure the integrity of embedded software systems, the activity of these systems must be monitored as they operate. For each of these systems, it is possible to establish a very succinct representation of nominal system activity. Furthermore, it is possible to detect departures from the nominal operating scenario in a timely fashion. Such departure may be due to various circumstances, e.g., an assault from an outside agent, thus forcing the system to operate in an off-nominal environment for which it was neither tested nor certified, or a hardware/software component that has ceased to operate in a nominal fashion. A well-designed system will have the property of graceful degradation. It must continue to run even though some of the functionality may have been lost. This involves the intelligent re-mapping of system functions. Those functions that are impacted by the failure of a system component must be identified and isolated. Thus, a system must be designed so that its basic operations may be re-mapped onto system components still operational. That is, the mission objectives of the software must be reassessed in terms of the current operational capabilities of the software system. By integrating the mechanisms to support observation and detection directly into the design methodology, we propose to shift

  20. Tank Monitoring and Document control System (TMACS) As Built Software Design Document

    Energy Technology Data Exchange (ETDEWEB)

    GLASSCOCK, J.A.

    2000-01-27

    This document describes the software design for the Tank Monitor and Control System (TMACS). This document captures the existing as-built design of TMACS as of November 1999. It will be used as a reference document to the system maintainers who will be maintaining and modifying the TMACS functions as necessary. The heart of the TMACS system is the ''point-processing'' functionality where a sample value is received from the field sensors and the value is analyzed, logged, or alarmed as required. This Software Design Document focuses on the point-processing functions.

  1. Green Software Engineering Adaption In Requirement Elicitation Process

    Directory of Open Access Journals (Sweden)

    Umma Khatuna Jannat

    2015-08-01

    Full Text Available A recent technology investigates the role of concern in the environment software that is green software system. Now it is widely accepted that the green software can fit all process of software development. It is also suitable for the requirement elicitation process. Now a days software companies have used requirements elicitation techniques in an enormous majority. Because this process plays more and more important roles in software development. At the present time most of the requirements elicitation process is improved by using some techniques and tools. So that the intention of this research suggests to adapt green software engineering for the intention of existing elicitation technique and recommend suitable actions for improvement. This research being involved qualitative data. I used few keywords in my searching procedure then searched IEEE ACM Springer Elsevier Google scholar Scopus and Wiley. Find out articles which published in 2010 until 2016. Finding from the literature review Identify 15 traditional requirement elicitations factors and 23 improvement techniques to convert green engineering. Lastly The paper includes a squat review of the literature a description of the grounded theory and some of the identity issues related finding of the necessity for requirements elicitation improvement techniques.

  2. Exploration on Automated Software Requirement Document Readability Approaches

    OpenAIRE

    Chen, Mingda; He, Yao

    2017-01-01

    Context. The requirements analysis phase, as the very beginning of software development process, has been identified as a quite important phase in the software development lifecycle. Software Requirement Specification (SRS) is the output of requirements analysis phase, whose quality factors play an important role in the evaluation work. Readability is a quite important SRS quality factor, but there are few available automated approaches for readability measurement, because of the tight depend...

  3. Information technologies in optimization process of monitoring of software and hardware status

    Science.gov (United States)

    Nikitin, P. V.; Savinov, A. N.; Bazhenov, R. I.; Ryabov, I. V.

    2018-05-01

    The article describes a model of a hardware and software monitoring system for a large company that provides customers with software as a service (SaaS solution) using information technology. The main functions of the monitoring system are: provision of up-todate data for analyzing the state of the IT infrastructure, rapid detection of the fault and its effective elimination. The main risks associated with the provision of these services are described; the comparative characteristics of the software are given; author's methods of monitoring the status of software and hardware are proposed.

  4. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    De Guio, Federico

    2013-01-01

    The Data Quality Monitoring (DQM) Software proved to be a central tool in the CMS experiment. Its flexibility allowed its integration in several environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release Validation, to constantly validate the functionality and the performance of the reconstruction software; in Monte Carlo productions. The central tool to deliver Data Quality information is a web site for browsing data quality histograms (DQM GUI). In this contribution the usage of the DQM Software in the different environments and its integration in the CMS Reconstruction Software Framework and in all production workflows are presented.

  5. Jagiellonian University Development of the LHCb VELO monitoring software platform

    CERN Document Server

    Majewski, Maciej

    2017-01-01

    One of the most important parts of the LHCb spectrometer is the VErtex LOcator (VELO), dedicated to the precise tracking close to the proton–proton interaction point. The quality of data produced by the VELO depends on the calibration process, which must be monitored to ensure its correctness. This work presents details on how the calibration monitoring is conducted and how it could be improved. It also includes information on monitoring software and data flow in the LHCb software framework.

  6. Software for marine ecological environment comprehensive monitoring system based on MCGS

    Science.gov (United States)

    Wang, X. H.; Ma, R.; Cao, X.; Cao, L.; Chu, D. Z.; Zhang, L.; Zhang, T. P.

    2017-08-01

    The automatic integrated monitoring software for marine ecological environment based on MCGS configuration software is designed and developed to realize real-time automatic monitoring of many marine ecological parameters. The DTU data transmission terminal performs network communication and transmits the data to the user data center in a timely manner. The software adopts the modular design and has the advantages of stable and flexible data structure, strong portability and scalability, clear interface, simple user operation and convenient maintenance. Continuous site comparison test of 6 months showed that, the relative error of the parameters monitored by the system such as temperature, salinity, turbidity, pH, dissolved oxygen was controlled within 5% with the standard method and the relative error of the nutrient parameters was within 15%. Meanwhile, the system had few maintenance times, low failure rate, stable and efficient continuous monitoring capabilities. The field application shows that the software is stable and the data communication is reliable, and it has a good application prospect in the field of marine ecological environment comprehensive monitoring.

  7. Light duty utility arm software requirements specification

    International Nuclear Information System (INIS)

    Kiebel, G.R.

    1995-01-01

    This document defines the software requirements for the integrated control and data acquisition system of the Light Duty Utility Arm (LDUA) System. It is intended to be used to guide the design of the application software, to be a basis for assessing the application software design, and to establish what is to be tested in the finished application software product

  8. 78 FR 47015 - Software Requirement Specifications for Digital Computer Software Used in Safety Systems of...

    Science.gov (United States)

    2013-08-02

    ... NUCLEAR REGULATORY COMMISSION [NRC-2012-0195] Software Requirement Specifications for Digital Computer Software Used in Safety Systems of Nuclear Power Plants AGENCY: Nuclear Regulatory Commission... issuing a revised regulatory guide (RG), revision 1 of RG 1.172, ``Software Requirement Specifications for...

  9. Section 508 Electronic Information Accessibility Requirements for Software Development

    Science.gov (United States)

    Ellis, Rebecca

    2014-01-01

    Section 508 Subpart B 1194.21 outlines requirements for operating system and software development in order to create a product that is accessible to users with various disabilities. This portion of Section 508 contains a variety of standards to enable those using assistive technology and with visual, hearing, cognitive and motor difficulties to access all information provided in software. The focus on requirements was limited to the Microsoft Windows® operating system as it is the predominant operating system used at this center. Compliance with this portion of the requirements can be obtained by integrating the requirements into the software development cycle early and by remediating issues in legacy software if possible. There are certain circumstances with software that may arise necessitating an exemption from these requirements, such as design or engineering software using dynamically changing graphics or numbers to convey information. These exceptions can be discussed with the Section 508 Coordinator and another method of accommodation used.

  10. Analyzing Software Requirements Errors in Safety-Critical, Embedded Systems

    Science.gov (United States)

    Lutz, Robyn R.

    1993-01-01

    This paper analyzes the root causes of safety-related software errors in safety-critical, embedded systems. The results show that software errors identified as potentially hazardous to the system tend to be produced by different error mechanisms than non- safety-related software errors. Safety-related software errors are shown to arise most commonly from (1) discrepancies between the documented requirements specifications and the requirements needed for correct functioning of the system and (2) misunderstandings of the software's interface with the rest of the system. The paper uses these results to identify methods by which requirements errors can be prevented. The goal is to reduce safety-related software errors and to enhance the safety of complex, embedded systems.

  11. Identify and Manage the Software Requirements Volatility

    OpenAIRE

    Khloud Abd Elwahab; Mahmoud Abd EL Latif; Sherif Kholeif

    2016-01-01

    Management of software requirements volatility through development of life cycle is a very important stage. It helps the team to control significant impact all over the project (cost, time and effort), and also it keeps the project on track, to finally satisfy the user which is the main success criteria for the software project. In this research paper, we have analysed the root causes of requirements volatility through a proposed framework presenting the requirements volatility causes and how...

  12. A software tool for design of process monitoring and analysis systems

    DEFF Research Database (Denmark)

    Singh, Ravendra; Gernaey, Krist; Gani, Rafiqul

    2009-01-01

    A well designed process monitoring and analysis system is necessary to consistently achieve any predefined end product quality. Systematic computer aided methods and tools provide the means to design the necessary process monitoring and analysis systems and/or to validate any existing monitoring...... and analysis system. A software to achieve this has been developed. Two developed supporting tools for the design, a knowledge base (consisting of the process knowledge as well as the knowledge on measurement methods & tools) and a model library (consisting of the process operational models) have been extended...... rigorously and integrated with the user interface, which made the software more generic and applicable to a wide range of problems. The software for the design of a process monitoring and analysis system is presented and illustrated with a tablet manufacturing process example....

  13. Proposing an Evidence-Based Strategy for Software Requirements Engineering.

    Science.gov (United States)

    Lindoerfer, Doris; Mansmann, Ulrich

    2016-01-01

    This paper discusses an evidence-based approach to software requirements engineering. The approach is called evidence-based, since it uses publications on the specific problem as a surrogate for stakeholder interests, to formulate risks and testing experiences. This complements the idea that agile software development models are more relevant, in which requirements and solutions evolve through collaboration between self-organizing cross-functional teams. The strategy is exemplified and applied to the development of a Software Requirements list used to develop software systems for patient registries.

  14. Capturing security requirements for software systems.

    Science.gov (United States)

    El-Hadary, Hassan; El-Kassas, Sherif

    2014-07-01

    Security is often an afterthought during software development. Realizing security early, especially in the requirement phase, is important so that security problems can be tackled early enough before going further in the process and avoid rework. A more effective approach for security requirement engineering is needed to provide a more systematic way for eliciting adequate security requirements. This paper proposes a methodology for security requirement elicitation based on problem frames. The methodology aims at early integration of security with software development. The main goal of the methodology is to assist developers elicit adequate security requirements in a more systematic way during the requirement engineering process. A security catalog, based on the problem frames, is constructed in order to help identifying security requirements with the aid of previous security knowledge. Abuse frames are used to model threats while security problem frames are used to model security requirements. We have made use of evaluation criteria to evaluate the resulting security requirements concentrating on conflicts identification among requirements. We have shown that more complete security requirements can be elicited by such methodology in addition to the assistance offered to developers to elicit security requirements in a more systematic way.

  15. Capturing security requirements for software systems

    Directory of Open Access Journals (Sweden)

    Hassan El-Hadary

    2014-07-01

    Full Text Available Security is often an afterthought during software development. Realizing security early, especially in the requirement phase, is important so that security problems can be tackled early enough before going further in the process and avoid rework. A more effective approach for security requirement engineering is needed to provide a more systematic way for eliciting adequate security requirements. This paper proposes a methodology for security requirement elicitation based on problem frames. The methodology aims at early integration of security with software development. The main goal of the methodology is to assist developers elicit adequate security requirements in a more systematic way during the requirement engineering process. A security catalog, based on the problem frames, is constructed in order to help identifying security requirements with the aid of previous security knowledge. Abuse frames are used to model threats while security problem frames are used to model security requirements. We have made use of evaluation criteria to evaluate the resulting security requirements concentrating on conflicts identification among requirements. We have shown that more complete security requirements can be elicited by such methodology in addition to the assistance offered to developers to elicit security requirements in a more systematic way.

  16. Capturing security requirements for software systems

    Science.gov (United States)

    El-Hadary, Hassan; El-Kassas, Sherif

    2014-01-01

    Security is often an afterthought during software development. Realizing security early, especially in the requirement phase, is important so that security problems can be tackled early enough before going further in the process and avoid rework. A more effective approach for security requirement engineering is needed to provide a more systematic way for eliciting adequate security requirements. This paper proposes a methodology for security requirement elicitation based on problem frames. The methodology aims at early integration of security with software development. The main goal of the methodology is to assist developers elicit adequate security requirements in a more systematic way during the requirement engineering process. A security catalog, based on the problem frames, is constructed in order to help identifying security requirements with the aid of previous security knowledge. Abuse frames are used to model threats while security problem frames are used to model security requirements. We have made use of evaluation criteria to evaluate the resulting security requirements concentrating on conflicts identification among requirements. We have shown that more complete security requirements can be elicited by such methodology in addition to the assistance offered to developers to elicit security requirements in a more systematic way. PMID:25685514

  17. Training Requirements and Information Management System. Software user guide

    Energy Technology Data Exchange (ETDEWEB)

    Cillan, T.F.; Hodgson, M.A.

    1992-05-01

    This is the software user`s guide for the Training Requirements and Information Management System. This guide defines and describes the software operating procedures as they apply to the end user of the software program. This guide is intended as a reference tool for the user who already has an indepth knowledge of the Training Requirements and Information Management System functions and data reporting requirement.

  18. Requirements engineering for software and systems

    CERN Document Server

    Laplante, Phillip A

    2014-01-01

    Solid requirements engineering has increasingly been recognized as the key to improved, on-time and on-budget delivery of software and systems projects. This book provides practical teaching for graduate and professional systems and software engineers. It uses extensive case studies and exercises to help students grasp concepts and techniques. With a focus on software-intensive systems, this text provides a probing and comprehensive review of recent developments in intelligent systems, soft computing techniques, and their diverse applications in manufacturing. The second edition contains 100% revised content and approximately 30% new material

  19. The Use of UML for Software Requirements Expression and Management

    Science.gov (United States)

    Murray, Alex; Clark, Ken

    2015-01-01

    It is common practice to write English-language "shall" statements to embody detailed software requirements in aerospace software applications. This paper explores the use of the UML language as a replacement for the English language for this purpose. Among the advantages offered by the Unified Modeling Language (UML) is a high degree of clarity and precision in the expression of domain concepts as well as architecture and design. Can this quality of UML be exploited for the definition of software requirements? While expressing logical behavior, interface characteristics, timeliness constraints, and other constraints on software using UML is commonly done and relatively straight-forward, achieving the additional aspects of the expression and management of software requirements that stakeholders expect, especially traceability, is far less so. These other characteristics, concerned with auditing and quality control, include the ability to trace a requirement to a parent requirement (which may well be an English "shall" statement), to trace a requirement to verification activities or scenarios which verify that requirement, and to trace a requirement to elements of the software design which implement that requirement. UML Use Cases, designed for capturing requirements, have not always been satisfactory. Some applications of them simply use the Use Case model element as a repository for English requirement statements. Other applications of Use Cases, in which Use Cases are incorporated into behavioral diagrams that successfully communicate the behaviors and constraints required of the software, do indeed take advantage of UML's clarity, but not in ways that support the traceability features mentioned above. Our approach uses the Stereotype construct of UML to precisely identify elements of UML constructs, especially behaviors such as State Machines and Activities, as requirements, and also to achieve the necessary mapping capabilities. We describe this approach in the

  20. Traceability of Requirements and Software Architecture for Change Management

    NARCIS (Netherlands)

    Göknil, Arda

    2011-01-01

    At the present day, software systems get more and more complex. The requirements of software systems change continuously and new requirements emerge frequently. New and/or modified requirements are integrated with the existing ones, and adaptations to the architecture and source code of the system

  1. SU-F-P-04: Implementation of Dose Monitoring Software: Successes and Pitfalls

    Energy Technology Data Exchange (ETDEWEB)

    Och, J [Geisinger Medical Center, Danville, PA (United States)

    2016-06-15

    Purpose: to successfully install a dose monitoring software (DMS) application to assist in CT protocol and dose management. Methods: Upon selecting the DMS, we began our implementation of the application. A working group composed of Medical Physics, Radiology Administration, Information Technology, and CT technologists was formed. On-site training in the application was supplied by the vendor. The decision was made to apply the process for all the CT protocols on all platforms at all facilities. Protocols were painstakingly mapped to the correct masters, and the system went ‘live’. Results: We are routinely using DMS as a tool in our Clinical Performance CT QA program. It is useful in determining the effectiveness of revisions to existing protocols, and establishing performance baselines for new units. However, the implementation was not without difficulty. We identified several pitfalls and obstacles which frustrated progress. Including: Training deficiencies, Nomenclature problems, Communication, DICOM variability. Conclusion: Dose monitoring software can be a potent tool for QA. However, implementation of the program can be problematic and requires planning, organization and commitment.

  2. Physics Simulation Software for Autonomous Propellant Loading and Gas House Autonomous System Monitoring

    Science.gov (United States)

    Regalado Reyes, Bjorn Constant

    2015-01-01

    1. Kennedy Space Center (KSC) is developing a mobile launching system with autonomous propellant loading capabilities for liquid-fueled rockets. An autonomous system will be responsible for monitoring and controlling the storage, loading and transferring of cryogenic propellants. The Physics Simulation Software will reproduce the sensor data seen during the delivery of cryogenic fluids including valve positions, pressures, temperatures and flow rates. The simulator will provide insight into the functionality of the propellant systems and demonstrate the effects of potential faults. This will provide verification of the communications protocols and the autonomous system control. 2. The High Pressure Gas Facility (HPGF) stores and distributes hydrogen, nitrogen, helium and high pressure air. The hydrogen and nitrogen are stored in cryogenic liquid state. The cryogenic fluids pose several hazards to operators and the storage and transfer equipment. Constant monitoring of pressures, temperatures and flow rates are required in order to maintain the safety of personnel and equipment during the handling and storage of these commodities. The Gas House Autonomous System Monitoring software will be responsible for constantly observing and recording sensor data, identifying and predicting faults and relaying hazard and operational information to the operators.

  3. Software requirements management based on use cases

    International Nuclear Information System (INIS)

    Xiao Jin

    2009-01-01

    In this paper, the requirements management based on use cases is theoretically explored, and a multi-layer use-case model is introduced, which combined with three levels of use cases and a single use-case refinement model. Through the practice in a software project, the multi-layer use-case model provides a good solution on how to control the requirements scope and change, and provides the balance of work assignment between customer departments, information management departments and software development outsourcing team. (authors)

  4. IMIS desktop & smartphone software solutions for monitoring spacecrafts' payload from anywhere

    Science.gov (United States)

    Baroukh, J.; Queyrut, O.; Airaud, J.

    In the past years, the demand for satellite remote operations has increased guided by on one hand, the will to reduce operations cost (on-call operators out of business hours), and on the other hand, the development of cooperation space missions resulting in a world wide distribution of engineers and science team members. Only a few off-the-shelf solutions exist to fulfill the need of remote payload monitoring, and they mainly use proprietary devices. The recent advent of mobile technologies (laptops, smartphones and tablets) as well as the worldwide deployment of broadband networks (3G, Wi-Fi hotspots), has opened up a technical window that brings new options. As part of the Mars Science Laboratory (MSL) mission, the Centre National D'Etudes Spatiales (CNES, the French space agency) has developed a new software solution for monitoring spacecraft payloads. The Instrument Monitoring Interactive Software (IMIS) offers state-of-the-art operational features for payload monitoring, and can be accessed remotely. It was conceived as a generic tool that can be used for heterogeneous payloads and missions. IMIS was designed as a classical client/server architecture. The server is hosted at CNES and acts as a data provider while two different kinds of clients are available depending on the level of mobility required. The first one is a rich client application, built on Eclipse framework, which can be installed on usual operating systems and communicates with the server through the Internet. The second one is a smartphone application for any Android platform, connected to the server thanks to the mobile broadband network or a Wi-Fi connection. This second client is mainly devoted to on-call operations and thus only contains a subset of the IMIS functionalities. This paper describes the operational context, including security aspects, that led IMIS development, presents the selected software architecture and details the various features of both clients: the desktop and the sm

  5. Traceability of Software Safety Requirements in Legacy Safety Critical Systems

    Science.gov (United States)

    Hill, Janice L.

    2007-01-01

    How can traceability of software safety requirements be created for legacy safety critical systems? Requirements in safety standards are imposed most times during contract negotiations. On the other hand, there are instances where safety standards are levied on legacy safety critical systems, some of which may be considered for reuse for new applications. Safety standards often specify that software development documentation include process-oriented and technical safety requirements, and also require that system and software safety analyses are performed supporting technical safety requirements implementation. So what can be done if the requisite documents for establishing and maintaining safety requirements traceability are not available?

  6. Light-Weight and Versatile Monitor for a Self-Adaptive Software Framework for IoT Systems

    Directory of Open Access Journals (Sweden)

    Young-Joo Kim

    2016-01-01

    Full Text Available Today, various Internet of Things (IoT devices and applications are being developed. Such IoT devices have different hardware (HW and software (SW capabilities; therefore, most applications require customization when IoT devices are changed or new applications are created. However, the applications executed on these devices are not optimized for power and performance because IoT device systems do not provide suitable static and dynamic information about fast-changing system resources and applications. Therefore, this paper proposes a light-weight and versatile monitor for a self-adaptive software framework to automatically control system resources according to the system status. The monitor helps running applications guarantee low power consumption and high performance for an optimal environment. The proposed monitor has two components: a monitoring component, which provides real-time static and dynamic information about system resources and applications, and a controlling component, which supports real-time control of system resources. For the experimental verification, we created a video transport system based on IoT devices and measured the CPU utilization by dynamic voltage and frequency scaling (DVFS for the monitor. The results demonstrate that, for up to 50 monitored processes, the monitor shows an average CPU utilization of approximately 4% in the three DVFS modes and demonstrates maximum optimization in the Performance mode of DVFS.

  7. A Method for Software Requirement Volatility Analysis Using QFD

    Directory of Open Access Journals (Sweden)

    Yunarso Anang

    2016-10-01

    Full Text Available Changes of software requirements are inevitable during the development life cycle. Rather than avoiding the circumstance, it is easier to just accept it and find a way to anticipate those changes. This paper proposes a method to analyze the volatility of requirement by using the Quality Function Deployment (QFD method and the introduced degree of volatility. Customer requirements are deployed to software functions and subsequently to architectural design elements. And then, after determining the potential for changes of the design elements, the degree of volatility of the software requirements is calculated. In this paper the method is described using a flow diagram and illustrated using a simple example, and is evaluated using a case study.

  8. Validation of a new software version for monitoring of the core of the Unit 2 of the Laguna Verde power plant with ARTS

    International Nuclear Information System (INIS)

    Calleros, G.; Riestra, M.; Ibanez, C.; Lopez, X.; Vargas, A.; Mendez, A.; Gomez, R.

    2005-01-01

    In this work it is intended a methodology to validate a new version of the software used for monitoring the reactor core, which requires of the evaluation of the thermal limits settled down in the Operation Technical Specifications, for the Unit 2 of Laguna Verde with ARTS (improvements to the APRMs, Rod Block Monitor and Technical specifications). According to the proposed methodology, those are shown differences found in the thermal limits determined with the new versions and previous of the core monitoring software. Author)

  9. The Node Monitoring Component of a Scalable Systems Software Environment

    Energy Technology Data Exchange (ETDEWEB)

    Miller, Samuel James [Iowa State Univ., Ames, IA (United States)

    2006-01-01

    This research describes Fountain, a suite of programs used to monitor the resources of a cluster. A cluster is a collection of individual computers that are connected via a high speed communication network. They are traditionally used by users who desire more resources, such as processing power and memory, than any single computer can provide. A common drawback to effectively utilizing such a large-scale system is the management infrastructure, which often does not often scale well as the system grows. Large-scale parallel systems provide new research challenges in the area of systems software, the programs or tools that manage the system from boot-up to running a parallel job. The approach presented in this thesis utilizes a collection of separate components that communicate with each other to achieve a common goal. While systems software comprises a broad array of components, this thesis focuses on the design choices for a node monitoring component. We will describe Fountain, an implementation of the Scalable Systems Software (SSS) node monitor specification. It is targeted at aggregate node monitoring for clusters, focusing on both scalability and fault tolerance as its design goals. It leverages widely used technologies such as XML and HTTP to present an interface to other components in the SSS environment.

  10. Computer software program for monitoring the availability of systems and components of electric power generating systems

    International Nuclear Information System (INIS)

    Petersen, T.A.; Hilsmeier, T.A.; Kapinus, D.M.

    1994-01-01

    As availabilities of electric power generating stations systems and components become more and more important from a financial, personnel safety, and regulatory requirements standpoint, it is evident that a comprehensive, yet simple and user-friendly program for system and component tracking and monitoring is needed to assist in effectively managing the large volume of systems and components with their large numbers of associated maintenance/availability records. A user-friendly computer software program for system and component availability monitoring has been developed that calculates, displays and monitors selected component and system availabilities. This is a Windows trademark based (Graphical User Interface) program that utilizes a system flow diagram for the data input screen which also provides a visual representation of availability values and limits for the individual components and associated systems. This program can be customized to the user's plant-specific system and component selections and configurations. As will be discussed herein, this software program is well suited for availability monitoring and ultimately providing valuable information for improving plant performance and reducing operating costs

  11. Toward an Agile Approach to Managing the Effect of Requirements on Software Architecture during Global Software Development

    OpenAIRE

    Alsahli, Abdulaziz; Khan, Hameed; Alyahya, Sultan

    2016-01-01

    Requirement change management (RCM) is a critical activity during software development because poor RCM results in occurrence of defects, thereby resulting in software failure. To achieve RCM, efficient impact analysis is mandatory. A common repository is a good approach to maintain changed requirements, reusing and reducing effort. Thus, a better approach is needed to tailor knowledge for better change management of requirements and architecture during global software development (GSD).The o...

  12. Tank monitor and control system (TMACS) software configuration management plan

    International Nuclear Information System (INIS)

    GLASSCOCK, J.A.

    1999-01-01

    This Software Configuration Management Plan (SCMP) describes the methodology for control of computer software developed and supported by the Systems Development and Integration (SD and I) organization of Lockheed Martin Services, Inc. (LMSI) for the Tank Monitor and Control System (TMACS). This plan controls changes to the software and configuration files used by TMACS. The controlled software includes the Gensym software package, Gensym knowledge base files developed for TMACS, C-language programs used by TMACS, the operating system on the production machine, language compilers, and all Windows NT commands and functions which affect the operating environment. The configuration files controlled include the files downloaded to the Acromag and Westronic field instruments

  13. Developing Control and Monitoring Software for the Data Acquisition System of the COMPASS Experiment at CERN

    Directory of Open Access Journals (Sweden)

    Martin Bodlák

    2013-01-01

    Full Text Available This paper focuses on the analysis, design and development of software for the new data acquisition system of the COMPASS experiment at CERN. In this system, the data flow is controlled by custom hardware; the software will therefore be used only for run control and for monitoring. The requirements on the software have been analyzed, and the functionality of the system has been defined. The system consists of several distributed nodes; communication between the nodes is based on a custom protocol and a DIM library. A minimal version of the system has already been implemented. Preliminary results of performance and stability tests have shown that the system fulfills the defined requirements, and is stable. In the next phase of development, the system will be tested on the real hardware. It is expected that the system will be ready for deployment in 2014.

  14. Data management for urban tree monitoring -- software requirements

    Science.gov (United States)

    Deborah J. ​Boyer; Lara A. Roman; Jason G. Henning; Matthew McFarland; Dana Dentice; Sarah C. Low; Casey Thomas; Glen. Abrams

    2016-01-01

    The creation of this report was organized by the Pennsylvania Horticultural Society (PHS) and the USDA Forest Service Philadelphia Field Station to explore how technology could be used to support the longterm systematic monitoring of urban trees by trained professionals, student interns and volunteers; assist with tree planting and maintenance data processes; and...

  15. Scorecard and KPIs for monitoring software factories effectiveness in the financial sector

    Directory of Open Access Journals (Sweden)

    Vicente Montequín

    2013-01-01

    Full Text Available Financial corporations and especially banking institutions have important needs concerning to the development of software around their business core. The software, that traditionally had been developed in house by the IT departments, is now usually outsourced to IT firms. These firms work under the software factories model. An important feature within this sector is that usually the financial groups keep the ownership of these firms because the strategic value of the software for the core business. These firms work almost exclusively for the owner financial group developing their software, but they have to demonstrate that they are so profitable and competitive like any other firm. The organizational structure of these firms has some differential features. Top level tasks (software design and project management are usually performed by the IT firm but the development is usually subcontracted to other software companies. Although financial corporations have always paid a special interest to investing in management and organizational policies to improve their efficiency, there have being always an important lack regarding to the control and monitoring of the software projects. They do not have suitable tools for monitoring actual process effectiveness. Adapting scorecards to this environment could be a useful tool for monitoring and improvement the process. Scorecard could here be used both as a tool for internal effectiveness measurement as well as externally, presenting sustainability indicators for the shareholders, the financial institutions. This paper aims to identify and define a collection of Key Performance Indicators which permit effectiveness to be improved under this context, focusing in the specific supply-chain model given by owner (financial group, software factory and software developers (subcontracted.

  16. Development of requirements tracking and verification technology for the NPP software

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Chul Hwan; Kim, Jang Yeol; Lee, Jang Soo; Song, Soon Ja; Lee, Dong Young; Kwon, Kee Choon

    1998-12-30

    Searched and analyzed the technology of requirements engineering in the areas of aerospace and defense industry, medical industry and nuclear industry. Summarized the status of tools for the software design and requirements management. Analyzed the software design methodology for the safety software of NPP. Development of the design requirements for the requirements tracking and verification system. Development of the background technology to design the prototype tool for the requirements tracking and verification.

  17. Development of requirements tracking and verification technology for the NPP software

    International Nuclear Information System (INIS)

    Jung, Chul Hwan; Kim, Jang Yeol; Lee, Jang Soo; Song, Soon Ja; Lee, Dong Young; Kwon, Kee Choon

    1998-01-01

    Searched and analyzed the technology of requirements engineering in the areas of aerospace and defense industry, medical industry and nuclear industry. Summarized the status of tools for the software design and requirements management. Analyzed the software design methodology for the safety software of NPP. Development of the design requirements for the requirements tracking and verification system. Development of the background technology to design the prototype tool for the requirements tracking and verification

  18. Enhancing requirements engineering for patient registry software systems with evidence-based components.

    Science.gov (United States)

    Lindoerfer, Doris; Mansmann, Ulrich

    2017-07-01

    Patient registries are instrumental for medical research. Often their structures are complex and their implementations use composite software systems to meet the wide spectrum of challenges. Commercial and open-source systems are available for registry implementation, but many research groups develop their own systems. Methodological approaches in the selection of software as well as the construction of proprietary systems are needed. We propose an evidence-based checklist, summarizing essential items for patient registry software systems (CIPROS), to accelerate the requirements engineering process. Requirements engineering activities for software systems follow traditional software requirements elicitation methods, general software requirements specification (SRS) templates, and standards. We performed a multistep procedure to develop a specific evidence-based CIPROS checklist: (1) A systematic literature review to build a comprehensive collection of technical concepts, (2) a qualitative content analysis to define a catalogue of relevant criteria, and (3) a checklist to construct a minimal appraisal standard. CIPROS is based on 64 publications and covers twelve sections with a total of 72 items. CIPROS also defines software requirements. Comparing CIPROS with traditional software requirements elicitation methods, SRS templates and standards show a broad consensus but differences in issues regarding registry-specific aspects. Using an evidence-based approach to requirements engineering for registry software adds aspects to the traditional methods and accelerates the software engineering process for registry software. The method we used to construct CIPROS serves as a potential template for creating evidence-based checklists in other fields. The CIPROS list supports developers in assessing requirements for existing systems and formulating requirements for their own systems, while strengthening the reporting of patient registry software system descriptions. It may be

  19. A SYSTEMATIC LITERATURE REVIEW ABOUT SOFTWARE REQUIREMENTS ELICITATION

    Directory of Open Access Journals (Sweden)

    LENIS R. WONG

    2017-02-01

    Full Text Available Requirements Elicitation is recognized as one of the most important activity in software development process as it has direct impact on its success. Although there are many proposals for improving this task, still there are issues which have to be solved. This paper aims to identify the current status of the latest researches related to software requirements elicitation through general framework for literature review, in order to answer the following research questions: Q1 What aspects have been covered by different proposal of requirements elicitation? Q2 What activities of the requirements elicitation process have been covered? And Q3 What factors influence on requirements elicitation and how? A cross-analysis of the outcome was performed. One of the results showed that requirements elicitation process needs improvements.

  20. Toward an Agile Approach to Managing the Effect of Requirements on Software Architecture during Global Software Development

    Directory of Open Access Journals (Sweden)

    Abdulaziz Alsahli

    2016-01-01

    Full Text Available Requirement change management (RCM is a critical activity during software development because poor RCM results in occurrence of defects, thereby resulting in software failure. To achieve RCM, efficient impact analysis is mandatory. A common repository is a good approach to maintain changed requirements, reusing and reducing effort. Thus, a better approach is needed to tailor knowledge for better change management of requirements and architecture during global software development (GSD.The objective of this research is to introduce an innovative approach for handling requirements and architecture changes simultaneously during global software development. The approach makes use of Case-Based Reasoning (CBR and agile practices. Agile practices make our approach iterative, whereas CBR stores requirements and makes them reusable. Twin Peaks is our base model, meaning that requirements and architecture are handled simultaneously. For this research, grounded theory has been applied; similarly, interviews from domain experts were conducted. Interview and literature transcripts formed the basis of data collection in grounded theory. Physical saturation of theory has been achieved through a published case study and developed tool. Expert reviews and statistical analysis have been used for evaluation. The proposed approach resulted in effective change management of requirements and architecture simultaneously during global software development.

  1. The software design of area γ radiation monitoring system

    International Nuclear Information System (INIS)

    Song Chenxin; Deng Changming; Cheng Chang; Ren Yi; Meng Dan; Liu Yun

    2007-01-01

    This paper main introduction the system structure, software architecture, design ideas of the area γ radiation monitoring system. Detailed introduction some programming technology about the computer communication with the local display unit. (authors)

  2. The software design of area γ radiation monitoring system

    International Nuclear Information System (INIS)

    Song Chenxin; Deng Changming; Cheng Chang; Ren Yi; Meng Dan; Liu Yun

    2008-01-01

    This paper main introduction the system structure, software architecture, design ideas of the area γ radiation monitoring system. Detailed introduction some programming technology about the computer communication with the local display unit. (authors)

  3. More about software requirements thorny issues and practical advice

    CERN Document Server

    Wiegers, Karl E

    2006-01-01

    No matter how much instruction you've had on managing software requirements, there's no substitute for experience. Too often, lessons about requirements engineering processes lack the no-nonsense guidance that supports real-world solutions. Complementing the best practices presented in his book, Software Requirements, Second Edition, requirements engineering authority Karl Wiegers tackles even more of the real issues head-on in this book. With straightforward, professional advice and practical solutions based on actual project experiences, this book answers many of the tough questions rais

  4. Tier-3 Monitoring Software Suite (T3MON) proposal

    CERN Document Server

    Andreeva, J; The ATLAS collaboration; Klimentov, A; Korenkov, V; Oleynik, D; Panitkin, S; Petrosyan, A

    2011-01-01

    The ATLAS Distributed Computing activities concentrated so far in the “central” part of the computing system of the experiment, namely the first 3 tiers (CERN Tier0, the 10 Tier1s centres and the 60+ Tier2s). This is a coherent system to perform data processing and management on a global scale and host (re)processing, simulation activities down to group and user analysis. Many ATLAS Institutes and National Communities built (or have plans to build) Tier-3 facilities. The definition of Tier-3 concept has been outlined (REFERENCE). Tier-3 centres consist of non-pledged resources mostly dedicated for the data analysis by the geographically close or local scientific groups. Tier-3 sites comprise a range of architectures and many do not possess Grid middleware, which would render application of Tier-2 monitoring systems useless. This document describes a strategy to develop a software suite for monitoring of the Tier3 sites. This software suite will enable local monitoring of the Tier3 sites and the global vie...

  5. Assessing students' performance in software requirements engineering education using scoring rubrics

    Science.gov (United States)

    Mkpojiogu, Emmanuel O. C.; Hussain, Azham

    2017-10-01

    The study investigates how helpful the use of scoring rubrics is, in the performance assessment of software requirements engineering students and whether its use can lead to students' performance improvement in the development of software requirements artifacts and models. Scoring rubrics were used by two instructors to assess the cognitive performance of a student in the design and development of software requirements artifacts. The study results indicate that the use of scoring rubrics is very helpful in objectively assessing the performance of software requirements or software engineering students. Furthermore, the results revealed that the use of scoring rubrics can also produce a good achievement assessments direction showing whether a student is either improving or not in a repeated or iterative assessment. In a nutshell, its use leads to the performance improvement of students. The results provided some insights for further investigation and will be beneficial to researchers, requirements engineers, system designers, developers and project managers.

  6. The Role of Requirements in the Success or Failure of Software Projects

    OpenAIRE

    Hussain, Azham; Mkpojiogu, Emmanuel O.C.; Kamal, Fazillah Mohmad

    2016-01-01

    Requirements engineering is pivotal and central to every successful software development project. There are several reasons why software projects fail; however, poorly elicited, documented, validated and managed requirements contribute grossly to software projects failure. Software project failures are normally very costly and risky and these could even a times be life threatening also. Projects that overlook requirements engineering processes often suffer or are most likely to suffer from fa...

  7. Development of solution monitoring software for enhanced safeguards at a large scale reprocessing facility

    Energy Technology Data Exchange (ETDEWEB)

    Van Handenhove, Carl; Breban, Domnica; Creusot, Christophe [International Atomic Energy Agency, Vienna (Austria); Dransart, Pascal; Dechamp, Luc [Joint Research Centre, European Commission, Ispra, Varese, (Italy); Jarde, Eric [Euriware, Equeurdreville (France)

    2011-12-15

    The implementation of an effective and efficient IAEA safeguards approach at large scale reprocessing facilities with large throughput and continuous flow of nuclear material requires the introduction of enhanced safeguards measures to provide added assurance about the absence of diversion of nuclear material and confirmation that the facility is operated as declared. One of the enhanced safeguards measures, a Solution Monitoring and Measurement System (SMMS), comprising data collection instruments, data transmission equipment and an advanced Solution Monitoring Software (SMS), is being implemented at a large scale reprocessing plant in Japan. SMS is designed as a tool to enable automatic calculations of volumes, densities and flow-rates in selected process vessels, including most of the vessels of the main nuclear material stream. This software also includes automatic features to support the inspectorate in verifying inventories and inventory changes. The software also enables one to analyze the flows of nuclear material within the process and of specified 'cycles' of operation, and, in order to provide assurance that the facility is being operated as declared to compare these with those expected (reference signatures). The configuration and parameterization work (especially the analytical and comparative work) for the implementation and configuration of the SMS has been carried out jointly between the IAEA, Euriware-France (the software developer) and the Joint Research Centre (JRC)-Ispra. This paper describes the main features of the SMS, including the principles underlying the automatic analysis functionalities. It then focuses on the collaborative work performed by the JRC-Ispra, Euriware and the IAEA for the parameterization of the software (vessels and cycles of operation), including the current status and the future challenges.

  8. Tank monitor and control system (TMACS) software configuration management plan; TOPICAL

    International Nuclear Information System (INIS)

    GLASSCOCK, J.A.

    1999-01-01

    This Software Configuration Management Plan (SCMP) describes the methodology for control of computer software developed and supported by the Systems Development and Integration (SD and I) organization of Lockheed Martin Services, Inc. (LMSI) for the Tank Monitor and Control System (TMACS). This plan controls changes to the software and configuration files used by TMACS. The controlled software includes the Gensym software package, Gensym knowledge base files developed for TMACS, C-language programs used by TMACS, the operating system on the production machine, language compilers, and all Windows NT commands and functions which affect the operating environment. The configuration files controlled include the files downloaded to the Acromag and Westronic field instruments

  9. Benchmarking therapeutic drug monitoring software: a review of available computer tools.

    Science.gov (United States)

    Fuchs, Aline; Csajka, Chantal; Thoma, Yann; Buclin, Thierry; Widmer, Nicolas

    2013-01-01

    Therapeutic drug monitoring (TDM) aims to optimize treatments by individualizing dosage regimens based on the measurement of blood concentrations. Dosage individualization to maintain concentrations within a target range requires pharmacokinetic and clinical capabilities. Bayesian calculations currently represent the gold standard TDM approach but require computation assistance. In recent decades computer programs have been developed to assist clinicians in this assignment. The aim of this survey was to assess and compare computer tools designed to support TDM clinical activities. The literature and the Internet were searched to identify software. All programs were tested on personal computers. Each program was scored against a standardized grid covering pharmacokinetic relevance, user friendliness, computing aspects, interfacing and storage. A weighting factor was applied to each criterion of the grid to account for its relative importance. To assess the robustness of the software, six representative clinical vignettes were processed through each of them. Altogether, 12 software tools were identified, tested and ranked, representing a comprehensive review of the available software. Numbers of drugs handled by the software vary widely (from two to 180), and eight programs offer users the possibility of adding new drug models based on population pharmacokinetic analyses. Bayesian computation to predict dosage adaptation from blood concentration (a posteriori adjustment) is performed by ten tools, while nine are also able to propose a priori dosage regimens, based only on individual patient covariates such as age, sex and bodyweight. Among those applying Bayesian calculation, MM-USC*PACK© uses the non-parametric approach. The top two programs emerging from this benchmark were MwPharm© and TCIWorks. Most other programs evaluated had good potential while being less sophisticated or less user friendly. Programs vary in complexity and might not fit all healthcare

  10. A Framework for Effective Software Monitoring in Project Management

    African Journals Online (AJOL)

    A Framework for Effective Software Monitoring in Project Management. ... is shown to provide meaningful interpretation of collected metric data by embedding certain quality function. Key words: Project Management, Feedback, project control, metrics, process model, quantitative validity ... AJOL African Journals Online.

  11. Monitoring and Discovery for Self-Organized Network Management in Virtualized and Software Defined Networks

    Science.gov (United States)

    Valdivieso Caraguay, Ángel Leonardo; García Villalba, Luis Javier

    2017-01-01

    This paper presents the Monitoring and Discovery Framework of the Self-Organized Network Management in Virtualized and Software Defined Networks SELFNET project. This design takes into account the scalability and flexibility requirements needed by 5G infrastructures. In this context, the present framework focuses on gathering and storing the information (low-level metrics) related to physical and virtual devices, cloud environments, flow metrics, SDN traffic and sensors. Similarly, it provides the monitoring data as a generic information source in order to allow the correlation and aggregation tasks. Our design enables the collection and storing of information provided by all the underlying SELFNET sublayers, including the dynamically onboarded and instantiated SDN/NFV Apps, also known as SELFNET sensors. PMID:28362346

  12. Monitoring and Discovery for Self-Organized Network Management in Virtualized and Software Defined Networks.

    Science.gov (United States)

    Caraguay, Ángel Leonardo Valdivieso; Villalba, Luis Javier García

    2017-03-31

    This paper presents the Monitoring and Discovery Framework of the Self-Organized Network Management in Virtualized and Software Defined Networks SELFNET project. This design takes into account the scalability and flexibility requirements needed by 5G infrastructures. In this context, the present framework focuses on gathering and storing the information (low-level metrics) related to physical and virtual devices, cloud environments, flow metrics, SDN traffic and sensors. Similarly, it provides the monitoring data as a generic information source in order to allow the correlation and aggregation tasks. Our design enables the collection and storing of information provided by all the underlying SELFNET sublayers, including the dynamically onboarded and instantiated SDN/NFV Apps, also known as SELFNET sensors.

  13. Software Defined Network Monitoring Scheme Using Spectral Graph Theory and Phantom Nodes

    Science.gov (United States)

    2014-09-01

    networks is the emergence of software - defined networking ( SDN ) [1]. SDN has existed for the...Chapter III for network monitoring. A. SOFTWARE DEFINED NETWORKS SDNs provide a new and innovative method to simplify network hardware by logically...and R. Giladi, “Performance analysis of software - defined networking ( SDN ),” in Proc. of IEEE 21st International Symposium on Modeling, Analysis

  14. TMS communications software. Volume 1: Computer interfaces

    Science.gov (United States)

    Brown, J. S.; Lenker, M. D.

    1979-01-01

    A prototype bus communications system, which is being used to support the Trend Monitoring System (TMS) as well as for evaluation of the bus concept is considered. Hardware and software interfaces to the MODCOMP and NOVA minicomputers are included. The system software required to drive the interfaces in each TMS computer is described. Documentation of other software for bus statistics monitoring and for transferring files across the bus is also included.

  15. Application Reuse Library for Software, Requirements, and Guidelines

    Science.gov (United States)

    Malin, Jane T.; Thronesbery, Carroll

    1994-01-01

    Better designs are needed for expert systems and other operations automation software, for more reliable, usable and effective human support. A prototype computer-aided Application Reuse Library shows feasibility of supporting concurrent development and improvement of advanced software by users, analysts, software developers, and human-computer interaction experts. Such a library expedites development of quality software, by providing working, documented examples, which support understanding, modification and reuse of requirements as well as code. It explicitly documents and implicitly embodies design guidelines, standards and conventions. The Application Reuse Library provides application modules with Demo-and-Tester elements. Developers and users can evaluate applicability of a library module and test modifications, by running it interactively. Sub-modules provide application code and displays and controls. The library supports software modification and reuse, by providing alternative versions of application and display functionality. Information about human support and display requirements is provided, so that modifications will conform to guidelines. The library supports entry of new application modules from developers throughout an organization. Example library modules include a timer, some buttons and special fonts, and a real-time data interface program. The library prototype is implemented in the object-oriented G2 environment for developing real-time expert systems.

  16. Nuclear power plant monitoring and control system software: verification and validation

    International Nuclear Information System (INIS)

    Kaneda, M.; Niki, K.; Shibata, K.

    1986-01-01

    The design philosophy, configuration, and production of process computer system software used for the monitoring and control of nuclear power plants are presented in detail. To achieve a very complex software system that not only has excellent performance, high reliability, and full fail safe protection, but also is easy to produce, verify, and validate, and has flexibility for future modifications, we developed the following software production system to support safe operation of nuclear power stations. The fundamental design philosophy of our monitoring and control system software is the complete separation of program logic from the data base. The logic section is highly standardized and applicable to a wide range of power generation plant computer application systems. The plant-unique properties and characteristics are all described in the data base. This separation of logic and data base has a dramatic effect on the reliability and productivity of the software system. One of the main features of the data base system is the use of easy-to-learn, easy-to-use, problem-oriented language that enables non-programmers to build up the data base using simple fill-in-the-blank type tables. The generation of these tables is fully automated, and the full set of online table editing utility software, which runs on the target plant process computer, has proven very effective in incorporation of changes and modifications at the site. (author)

  17. New software tool for dynamic radiological characterisation and monitoring in nuclear sites

    International Nuclear Information System (INIS)

    Szoeke, Istvan; Louka, Michael N.; Mark, Niels K.; Bryntesen, Tom R.; Bratteli, Joachim; Edvardsen, Svein T.; Gustavsen, Morten A.; Toppe, Aleksander L.; Johnsen, Terje; Rindahl, Grete

    2012-01-01

    The Halden Reactor Project (HRP) is a jointly sponsored international cooperation, under the aegis of the Organisation for Economic Co-operation and Development - Nuclear Energy Agency. Extensive and valuable guidance and tools, connected to safe and reliable operation of nuclear facilities, has been elaborated throughout the years within the frame of this programme. The HRP has particularly high level results in virtual-reality based tools for real-time areal and personal monitoring. The techniques, developed earlier, are now being supplemented to enhance the planning and monitoring capabilities, and support general radiological characterisation connected to nuclear sites and facilities. Due to the complexity and abundance of the input information required, software tools, dedicated to the radiological characterization of contaminated materials, buildings, land and groundwater, are applied to review, evaluate and visualize the data. Characterisation of the radiation situation in a realistic environment can be very complex, and efficient visualisation of the data to the user is not straight forward. The monitoring and planning tools elaborated in the frame of the HRP feature very sophisticated three-dimensional (3D) high definition visualisation and user interfaces to promote easy interpretation of the input data. The visualisation tools permit dynamic visualisation of radiation fields in virtual or augmented reality by various techniques and real-time personal monitoring of humanoid models. In addition new techniques are being elaborated to visualise the 3D distribution of activities in structures and materials. The dosimetric algorithms, feeding information to the visualisation and user interface of these planning tools, include deterministic radiation transport techniques suitable for fast photon dose estimates, in case physical and radio- and spectrometric characteristics of the gamma sources are known. The basic deterministic model, implemented in earlier

  18. Space shuttle orbiter guidance, naviagation and control software functional requirements: Horizontal flight operations

    Science.gov (United States)

    1972-01-01

    The shuttle GN&C software functions for horizontal flight operations are defined. Software functional requirements are grouped into two categories: first horizontal flight requirements and full mission horizontal flight requirements. The document privides the intial step in the shuttle GN&C software design process. It also serves as a management tool to identify analyses which are required to define requirements.

  19. Evaluation procedure of software requirements specification for digital I and C of KNGR

    International Nuclear Information System (INIS)

    Lee, Jang Soo; Park, Jong Kyun; Lee, Ki Young; Kim, Jang Yeol; Cheon, Se Woo

    2001-06-01

    The accuracy of the specification of requirements of a digital system is of prime importance to the acceptance and success of the system. The development, use, and regulation of computer systems in nuclear reactor Instrumentation and Control (I and C) systems to enhance reliability and safety is a complex issue. This report is one of a series of reports from the Korean Next Generation Reactor (KNGR) Software Safety Verification and Validation (SSVV) Task, Korea Atomic Energy Research Institute, which investigates different aspects of computer software in reactor I and C systems, and describes the engineering procedures for developing such a software. The purpose of this guideline is to give the software safety evaluator the trail map between the code and standards layer and the design methodology and documents layer for the software important to safety in nuclear power plants. Recently, the requirements specification of safety-critical software systems and safety analysis of them are being recognized as one of the important issues in the software life cycle, and being developed new regulatory positions and standards by the regulatory and the standardization organizations such as IAEA, IEC, and IEEE. We presented the procedure for evaluating the software requirements specifications of the KNGR protection systems. We believe it can be useful for both licenser and licensee to conduct an evaluation of the safety in the requirements phase of developing the software. The guideline consists of the requirements engineering for software of KNGR protection systems in chapter 1, the evaluation checklist of software requirements specification in chapter2.3, and the safety evaluation procedure of KNGR software requirements specification in chapter 2.4

  20. Software use cases to elicit the software requirements analysis within the ASTRI project

    Science.gov (United States)

    Conforti, Vito; Antolini, Elisa; Bonnoli, Giacomo; Bruno, Pietro; Bulgarelli, Andrea; Capalbi, Milvia; Fioretti, Valentina; Fugazza, Dino; Gardiol, Daniele; Grillo, Alessandro; Leto, Giuseppe; Lombardi, Saverio; Lucarelli, Fabrizio; Maccarone, Maria Concetta; Malaguti, Giuseppe; Pareschi, Giovanni; Russo, Federico; Sangiorgi, Pierluca; Schwarz, Joseph; Scuderi, Salvatore; Tanci, Claudio; Tosti, Gino; Trifoglio, Massimo; Vercellone, Stefano; Zanmar Sanchez, Ricardo

    2016-07-01

    The Italian National Institute for Astrophysics (INAF) is leading the Astrofisica con Specchi a Tecnologia Replicante Italiana (ASTRI) project whose main purpose is the realization of small size telescopes (SST) for the Cherenkov Telescope Array (CTA). The first goal of the ASTRI project has been the development and operation of an innovative end-to-end telescope prototype using a dual-mirror optical configuration (SST-2M) equipped with a camera based on silicon photo-multipliers and very fast read-out electronics. The ASTRI SST-2M prototype has been installed in Italy at the INAF "M.G. Fracastoro" Astronomical Station located at Serra La Nave, on Mount Etna, Sicily. This prototype will be used to test several mechanical, optical, control hardware and software solutions which will be used in the ASTRI mini-array, comprising nine telescopes proposed to be placed at the CTA southern site. The ASTRI mini-array is a collaborative and international effort led by INAF and carried out by Italy, Brazil and South-Africa. We present here the use cases, through UML (Unified Modeling Language) diagrams and text details, that describe the functional requirements of the software that will manage the ASTRI SST-2M prototype, and the lessons learned thanks to these activities. We intend to adopt the same approach for the Mini Array Software System that will manage the ASTRI miniarray operations. Use cases are of importance for the whole software life cycle; in particular they provide valuable support to the validation and verification activities. Following the iterative development approach, which breaks down the software development into smaller chunks, we have analysed the requirements, developed, and then tested the code in repeated cycles. The use case technique allowed us to formalize the problem through user stories that describe how the user procedurally interacts with the software system. Through the use cases we improved the communication among team members, fostered

  1. Monitoring the software quality in FairRoot

    Energy Technology Data Exchange (ETDEWEB)

    Uhlig, Florian; Al-Turany, Mohammad [GSI, Darmstadt (Germany)

    2010-07-01

    Up-to-date informations about a software project helps to find problems as early as possible. This includes for example information if a software project can be build on all supported platforms without errors or if specified tests can be executed and deliver the correct results. We present the scheme which is used within the FairRoot framework to continuously monitor the status of the project. The tools used for these tasks are based on the open source tools CMake and CDash. CMake is used to generate standard build files for the different operating systems/compiler out of simple configuration files and to steer the build and test processes. The generated information is send to a central CDash server. From the generated web pages information about the status of the project at any given time can be obtained.

  2. Monitoring and Discovery for Self-Organized Network Management in Virtualized and Software Defined Networks

    Directory of Open Access Journals (Sweden)

    Ángel Leonardo Valdivieso Caraguay

    2017-03-01

    Full Text Available This paper presents the Monitoring and Discovery Framework of the Self-Organized Network Management in Virtualized and Software Defined Networks SELFNET project. This design takes into account the scalability and flexibility requirements needed by 5G infrastructures. In this context, the present framework focuses on gathering and storing the information (low-level metrics related to physical and virtual devices, cloud environments, flow metrics, SDN traffic and sensors. Similarly, it provides the monitoring data as a generic information source in order to allow the correlation and aggregation tasks. Our design enables the collection and storing of information provided by all the underlying SELFNET sublayers, including the dynamically onboarded and instantiated SDN/NFV Apps, also known as SELFNET sensors.

  3. Generic Safety Requirements for Developing Safe Insulin Pump Software

    Science.gov (United States)

    Zhang, Yi; Jetley, Raoul; Jones, Paul L; Ray, Arnab

    2011-01-01

    Background The authors previously introduced a highly abstract generic insulin infusion pump (GIIP) model that identified common features and hazards shared by most insulin pumps on the market. The aim of this article is to extend our previous work on the GIIP model by articulating safety requirements that address the identified GIIP hazards. These safety requirements can be validated by manufacturers, and may ultimately serve as a safety reference for insulin pump software. Together, these two publications can serve as a basis for discussing insulin pump safety in the diabetes community. Methods In our previous work, we established a generic insulin pump architecture that abstracts functions common to many insulin pumps currently on the market and near-future pump designs. We then carried out a preliminary hazard analysis based on this architecture that included consultations with many domain experts. Further consultation with domain experts resulted in the safety requirements used in the modeling work presented in this article. Results Generic safety requirements for the GIIP model are presented, as appropriate, in parameterized format to accommodate clinical practices or specific insulin pump criteria important to safe device performance. Conclusions We believe that there is considerable value in having the diabetes, academic, and manufacturing communities consider and discuss these generic safety requirements. We hope that the communities will extend and revise them, make them more representative and comprehensive, experiment with them, and use them as a means for assessing the safety of insulin pump software designs. One potential use of these requirements is to integrate them into model-based engineering (MBE) software development methods. We believe, based on our experiences, that implementing safety requirements using MBE methods holds promise in reducing design/implementation flaws in insulin pump development and evolutionary processes, therefore improving

  4. Coupling Sensing Hardware with Data Interrogation Software for Structural Health Monitoring

    Directory of Open Access Journals (Sweden)

    Charles R. Farrar

    2006-01-01

    Full Text Available The process of implementing a damage detection strategy for aerospace, civil and mechanical engineering infrastructure is referred to as structural health monitoring (SHM. The authors' approach is to address the SHM problem in the context of a statistical pattern recognition paradigm. In this paradigm, the process can be broken down into four parts: (1 Operational Evaluation, (2 Data Acquisition and Cleansing, (3 Feature Extraction and Data Compression, and (4 Statistical Model Development for Feature Discrimination. These processes must be implemented through hardware or software and, in general, some combination of these two approaches will be used. This paper will discuss each portion of the SHM process with particular emphasis on the coupling of a general purpose data interrogation software package for structural health monitoring with a modular wireless sensing and processing platform. More specifically, this paper will address the need to take an integrated hardware/software approach to developing SHM solutions.

  5. Spacelab user implementation assessment study. (Software requirements analysis). Volume 2: Technical report

    Science.gov (United States)

    1976-01-01

    The engineering analyses and evaluation studies conducted for the Software Requirements Analysis are discussed. Included are the development of the study data base, synthesis of implementation approaches for software required by both mandatory onboard computer services and command/control functions, and identification and implementation of software for ground processing activities.

  6. Belle II Software

    International Nuclear Information System (INIS)

    Kuhr, T; Ritter, M

    2016-01-01

    Belle II is a next generation B factory experiment that will collect 50 times more data than its predecessor, Belle. The higher luminosity at the SuperKEKB accelerator leads to higher background levels and requires a major upgrade of the detector. As a consequence, the simulation, reconstruction, and analysis software must also be upgraded substantially. Most of the software has been redesigned from scratch, taking into account the experience from Belle and other experiments and utilizing new technologies. The large amount of experimental and simulated data requires a high level of reliability and reproducibility, even in parallel environments. Several technologies, tools, and organizational measures are employed to evaluate and monitor the performance of the software during development. (paper)

  7. Knowledge-based software design for Defense-in-Depth risk monitor system and application for AP1000

    International Nuclear Information System (INIS)

    Ma Zhanguo; Yoshikawa, Hidekazu; Yang Ming; Nakagawa, Takashi

    2017-01-01

    As part of the new risk monitor system, the software for the plant Defense-in-Depth (DiD) risk monitor system was designed based on the state-transition and finite-state machine, and then the knowledge-based software was developed by object-oriented method utilizing the Unified Modeling Language (UML). Currently, there are mainly two functions in the developed plant DiD risk monitor software that are knowledge-base editor which is used to model the system in a hierarchical manner and the interaction simulator that simulates the interactions between the different actors in the model. In this paper, a model for playing its behavior is called an Actor which is modeled at the top level. The passive safety AP1000 power plant was studied and the small-break loss-of-coolant accident (SBLOCA) design basis accident transient is modeled using the plant DiD risk monitor software. Furthermore, the simulation result is shown for the interactions between the actors which are defined in the plant DiD risk monitor system as PLANT actor, OPERATOR actor, and SUPERVISOR actor. This paper shows that it is feasible to model the nuclear power plant knowledge base using the software modeling technique. The software can make the large knowledge base for the nuclear power plant with small effort. (author)

  8. Surveillance Analysis Computer System (SACS): Software requirements specification (SRS). Revision 2

    International Nuclear Information System (INIS)

    Glasscock, J.A.

    1995-01-01

    This document is the primary document establishing requirements for the Surveillance Analysis Computer System (SACS) database, an Impact Level 3Q system. SACS stores information on tank temperatures, surface levels, and interstitial liquid levels. This information is retrieved by the customer through a PC-based interface and is then available to a number of other software tools. The software requirements specification (SRS) describes the system requirements for the SACS Project, and follows the Standard Engineering Practices (WHC-CM-6-1), Software Practices (WHC-CM-3-10) and Quality Assurance (WHC-CM-4-2, QR 19.0) policies

  9. Control, Test and Monitoring Software Framework for the ATLAS Level-1 Calorimeter Trigger

    CERN Document Server

    Achenbach, R; Aharrouche, M; Andrei, V; Åsman, B; Barnett, B M; Bauss, B; Bendel, M; Bohm, C; Booth, J R A; Bracinik, J; Brawn, I P; Charlton, D G; Childers, J T; Collins, N J; Curtis, C J; Davis, A O; Eckweiler, S; Eisenhandler, E F; Faulkner, P J W; Fleckner, J; Föhlisch, F; Gee, C N P; Gillman, A R; Goringer, C; Groll, M; Hadley, D R; Hanke, P; Hellman, S; Hidvegi, A; Hillier, S J; Johansen, M; Kluge, E E; Kühl, T; Landon, M; Lendermann, V; Lilley, J N; Mahboubi, K; Mahout, G; Meier, K; Middleton, R P; Moa, T; Morris, J D; Müller, F; Neusiedl, A; Ohm, C; Oltmann, B; Perera, V J O; Prieur, D P F; Qian, W; Rieke, S; Rühr, F; Sankey, D P C; Schäfer, U; Schmitt, K; Schultz-Coulon, H C; Silverstein, S; Sjölin, J; Staley, R J; Stamen, R; Stockton, M C; Tan, C L A; Tapprogge, S; Thomas, J P; Thompson, P D; Watkins, P M; Watson, A; Weber, P; Wessels, M; Wildt, M

    2008-01-01

    The ATLAS first-level calorimeter trigger is a hardware-based system designed to identify high-pT jets, electron/photon and tau candidates and to measure total and missing ET in the ATLAS calorimeters. The complete trigger system consists of over 300 customdesignedVME modules of varying complexity. These modules are based around FPGAs or ASICs with many configurable parameters, both to initialize the system with correct calibrations and timings and to allow flexibility in the trigger algorithms. The control, testing and monitoring of these modules requires a comprehensive, but well-designed and modular, software framework, which we will describe in this paper.

  10. iAssist: a software framework for intelligent patient monitoring.

    Science.gov (United States)

    Brouse, Christopher; Dumont, Guy; Yang, Ping; Lim, Joanne; Ansermino, J Mark

    2007-01-01

    A software framework (iAssist) has been developed for intelligent patient monitoring, and forms the foundation of a clinical monitoring expert system. The framework is extensible, flexible, and interoperable. It supports plugins to perform data acquisition, signal processing, graphical display, data storage, and output to external devices. iAssist currently incorporates two plugins to detect change point events in physiological trends. In 38 surgical cases, iAssist detected 868 events, of which clinicians rated more than 50% as clinically significant and less than 7% as artifacts. Clinicians found iAssist intuitive and easy to use.

  11. Software Testing Requires Variability

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak

    2003-01-01

    Software variability is the ability of a software system or artefact to be changed, customized or configured for use in a particular context. Variability in software systems is important from a number of perspectives. Some perspectives rightly receive much attention due to their direct economic...... impact in software production. As is also apparent from the call for papers these perspectives focus on qualities such as reuse, adaptability, and maintainability....

  12. SIMPATIQCO: a server-based software suite which facilitates monitoring the time course of LC-MS performance metrics on Orbitrap instruments.

    Science.gov (United States)

    Pichler, Peter; Mazanek, Michael; Dusberger, Frederico; Weilnböck, Lisa; Huber, Christian G; Stingl, Christoph; Luider, Theo M; Straube, Werner L; Köcher, Thomas; Mechtler, Karl

    2012-11-02

    While the performance of liquid chromatography (LC) and mass spectrometry (MS) instrumentation continues to increase, applications such as analyses of complete or near-complete proteomes and quantitative studies require constant and optimal system performance. For this reason, research laboratories and core facilities alike are recommended to implement quality control (QC) measures as part of their routine workflows. Many laboratories perform sporadic quality control checks. However, successive and systematic longitudinal monitoring of system performance would be facilitated by dedicated automatic or semiautomatic software solutions that aid an effortless analysis and display of QC metrics over time. We present the software package SIMPATIQCO (SIMPle AuTomatIc Quality COntrol) designed for evaluation of data from LTQ Orbitrap, Q-Exactive, LTQ FT, and LTQ instruments. A centralized SIMPATIQCO server can process QC data from multiple instruments. The software calculates QC metrics supervising every step of data acquisition from LC and electrospray to MS. For each QC metric the software learns the range indicating adequate system performance from the uploaded data using robust statistics. Results are stored in a database and can be displayed in a comfortable manner from any computer in the laboratory via a web browser. QC data can be monitored for individual LC runs as well as plotted over time. SIMPATIQCO thus assists the longitudinal monitoring of important QC metrics such as peptide elution times, peak widths, intensities, total ion current (TIC) as well as sensitivity, and overall LC-MS system performance; in this way the software also helps identify potential problems. The SIMPATIQCO software package is available free of charge.

  13. Validation of a new software version for monitoring of the core of the Unit 2 of the Laguna Verde power plant with ARTS; Validacion de una nueva version del software para monitoreo del nucleo de la Unidad 2 de la Central Laguna Verde con ARTS

    Energy Technology Data Exchange (ETDEWEB)

    Calleros, G.; Riestra, M.; Ibanez, C.; Lopez, X.; Vargas, A.; Mendez, A.; Gomez, R. [CFE, Central Nucleoelectrica de Laguna Verde, Alto Lucero, Veracruz (Mexico)]. e-mail: gcm9acpp@cfe.gob.mx

    2005-07-01

    In this work it is intended a methodology to validate a new version of the software used for monitoring the reactor core, which requires of the evaluation of the thermal limits settled down in the Operation Technical Specifications, for the Unit 2 of Laguna Verde with ARTS (improvements to the APRMs, Rod Block Monitor and Technical specifications). According to the proposed methodology, those are shown differences found in the thermal limits determined with the new versions and previous of the core monitoring software. Author)

  14. Quality assurance requirements for the computer software and safety analyses

    International Nuclear Information System (INIS)

    Husarecek, J.

    1992-01-01

    The requirements are given as placed on the development, procurement, maintenance, and application of software for the creation or processing of data during the design, construction, operation, repair, maintenance and safety-related upgrading of nuclear power plants. The verification and validation processes are highlighted, and the requirements put on the software documentation are outlined. The general quality assurance principles applied to safety analyses are characterized. (J.B.). 1 ref

  15. IAEA/NDS requirements related to database software

    International Nuclear Information System (INIS)

    Pronyaev, V.; Zerkin, V.

    2001-01-01

    Full text: The Nuclear Data Section of the IAEA disseminates data to the NDS users through Internet or on CD-ROMs and diskettes. OSU Web-server on DEC Alpha with Open VMS and Oracle/DEC DBMS provides via CGI scripts and FORTRAN retrieval programs access to the main nuclear databases supported by the networks of Nuclear Reactions Data Centres and Nuclear Structure and Decay Data Centres (CINDA, EXFOR, ENDF, NSR, ENSDF). For Web-access to data from other libraries and files, hyper-links to the files stored in ASCII text or other formats are used. Databases on CD-ROM are usually provided with some retrieval system. They are distributed in the run-time mode and comply with all license requirements for software used in their development. Although major development work is done now at the PC with MS-Windows and Linux, NDS may not at present, due to some institutional conditions, use these platforms for organization of the Web access to the data. Starting the end of 1999, the NDS, in co-operation with other data centers, began to work out the strategy of migration of main network nuclear data bases onto platforms other than DEC Alpha/Open VMS/DBMS. Because the different co-operating centers have their own preferences for hardware and software, the requirement to provide maximum platform independence for nuclear databases is the most important and desirable feature. This requirement determined some standards for the nuclear database software development. Taking into account the present state and future development, these standards can be formulated as follows: 1. All numerical data (experimental, evaluated, recommended values and their uncertainties) prepared for inclusion in the IAEA/NDS nuclear database should be submitted in the form of the ASCII text files and will be kept at NDS as a master file. 2. Databases with complex structure should be submitted in the form of the files with standard SQL statements describing all its components. All extensions of standard SQL

  16. Software safety analysis on the model specified by NuSCR and SMV input language at requirements phase of software development life cycle using SMV

    International Nuclear Information System (INIS)

    Koh, Kwang Yong; Seong, Poong Hyun

    2005-01-01

    Safety-critical software process is composed of development process, verification and validation (V and V) process and safety analysis process. Safety analysis process has been often treated as an additional process and not found in a conventional software process. But software safety analysis (SSA) is required if software is applied to a safety system, and the SSA shall be performed independently for the safety software through software development life cycle (SDLC). Of all the phases in software development, requirements engineering is generally considered to play the most critical role in determining the overall software quality. NASA data demonstrate that nearly 75% of failures found in operational software were caused by errors in the requirements. The verification process in requirements phase checks the correctness of software requirements specification, and the safety analysis process analyzes the safety-related properties in detail. In this paper, the method for safety analysis at requirements phase of software development life cycle using symbolic model verifier (SMV) is proposed. Hazard is discovered by hazard analysis and in other to use SMV for the safety analysis, the safety-related properties are expressed by computation tree logic (CTL)

  17. Requirements on software lifecycle process (RSLP) for KALIMER digital computer-based MMIS design

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Jang Soo; Kwon, Kee Choon; Kim, Jang Yeol [Korea Atomic Energy Research Institute, Taejon (Korea)

    1998-04-01

    Digital Man Machine Interface System (MMIS) systems of Korea Advanced Liquid MEtal Reactor (KALIMER) may share code, data transmission, data, and process equipment to a greater degree than analog systems. Although this sharing is the basis for many of the advantages of digital systems, it also raises a key concern: a design using shared data or code has the potential to propagate a common-cause or common-mode failure via software errors, thus defeating the redundancy achieved by the hardware architectural structure. Greater sharing of process equipment among functions within a channel increases the consequences of the failure of a single hardware module and reduces the amount of diversity available within a single safety channel. The software safety plan describes the safety analysis implementation tasks that are to be carried out during the software life cycle. Documentation should exist that shows that the safety analysis activities have been successfully accomplished for each life cycle activity group. In particular, the documentation should show that the system safety requirement have been adequately addressed for each life cycle activity group, that no new hazards have been introduced, and that the software requirements, design elements, and code elements that can affect safety have been identified. Because the safety of software can be assured through both the process Verification and Validation (V and V) itself and the V and V of all the intermediate and final products during the software development lifecycle, the development of KALIMER Software Safety Framework (KSSF) must be established. As the first activity for establishing KSSF, we have developed this report, Requirement on Software Life-cycle Process (RSLP) for designing KALIMER digital MMIS. This report is organized as follows. Section I describes the background, definitions, and references of RSLP. Section II describes KALIMER safety software categorization. In Section III, we define the

  18. WE-D-9A-06: Open Source Monitor Calibration and Quality Control Software for Enterprise Display Management

    Energy Technology Data Exchange (ETDEWEB)

    Bevins, N; Vanderhoek, M; Lang, S; Flynn, M [Henry Ford Health System, Detroit, MI (United States)

    2014-06-15

    Purpose: Medical display monitor calibration and quality control present challenges to medical physicists. The purpose of this work is to demonstrate and share experiences with an open source package that allows for both initial monitor setup and routine performance evaluation. Methods: A software package, pacsDisplay, has been developed over the last decade to aid in the calibration of all monitors within the radiology group in our health system. The software is used to calibrate monitors to follow the DICOM Grayscale Standard Display Function (GSDF) via lookup tables installed on the workstation. Additional functionality facilitates periodic evaluations of both primary and secondary medical monitors to ensure satisfactory performance. This software is installed on all radiology workstations, and can also be run as a stand-alone tool from a USB disk. Recently, a database has been developed to store and centralize the monitor performance data and to provide long-term trends for compliance with internal standards and various accrediting organizations. Results: Implementation and utilization of pacsDisplay has resulted in improved monitor performance across the health system. Monitor testing is now performed at regular intervals and the software is being used across multiple imaging modalities. Monitor performance characteristics such as maximum and minimum luminance, ambient luminance and illuminance, color tracking, and GSDF conformity are loaded into a centralized database for system performance comparisons. Compliance reports for organizations such as MQSA, ACR, and TJC are generated automatically and stored in the same database. Conclusion: An open source software solution has simplified and improved the standardization of displays within our health system. This work serves as an example method for calibrating and testing monitors within an enterprise health system.

  19. WE-D-9A-06: Open Source Monitor Calibration and Quality Control Software for Enterprise Display Management

    International Nuclear Information System (INIS)

    Bevins, N; Vanderhoek, M; Lang, S; Flynn, M

    2014-01-01

    Purpose: Medical display monitor calibration and quality control present challenges to medical physicists. The purpose of this work is to demonstrate and share experiences with an open source package that allows for both initial monitor setup and routine performance evaluation. Methods: A software package, pacsDisplay, has been developed over the last decade to aid in the calibration of all monitors within the radiology group in our health system. The software is used to calibrate monitors to follow the DICOM Grayscale Standard Display Function (GSDF) via lookup tables installed on the workstation. Additional functionality facilitates periodic evaluations of both primary and secondary medical monitors to ensure satisfactory performance. This software is installed on all radiology workstations, and can also be run as a stand-alone tool from a USB disk. Recently, a database has been developed to store and centralize the monitor performance data and to provide long-term trends for compliance with internal standards and various accrediting organizations. Results: Implementation and utilization of pacsDisplay has resulted in improved monitor performance across the health system. Monitor testing is now performed at regular intervals and the software is being used across multiple imaging modalities. Monitor performance characteristics such as maximum and minimum luminance, ambient luminance and illuminance, color tracking, and GSDF conformity are loaded into a centralized database for system performance comparisons. Compliance reports for organizations such as MQSA, ACR, and TJC are generated automatically and stored in the same database. Conclusion: An open source software solution has simplified and improved the standardization of displays within our health system. This work serves as an example method for calibrating and testing monitors within an enterprise health system

  20. Solid Waste Information and Tracking System (SWITS) Software Requirements Specification

    International Nuclear Information System (INIS)

    MAY, D.L.

    2000-01-01

    This document is the primary document establishing requirements for the Solid Waste Information and Tracking System (SWITS) as it is converted to a client-server architecture. The purpose is to provide the customer and the performing organizations with the requirements for the SWITS in the new environment. This Software Requirement Specification (SRS) describes the system requirements for the SWITS Project, and follows the PHMC Engineering Requirements, HNF-PRO-1819, and Computer Software Qualify Assurance Requirements, HNF-PRO-309, policies. This SRS includes sections on general description, specific requirements, references, appendices, and index. The SWITS system defined in this document stores information about the solid waste inventory on the Hanford site. Waste is tracked as it is generated, analyzed, shipped, stored, and treated. In addition to inventory reports a number of reports for regulatory agencies are produced

  1. Solid Waste Information and Tracking System (SWITS) Software Requirements Specification

    Energy Technology Data Exchange (ETDEWEB)

    MAY, D.L.

    2000-03-22

    This document is the primary document establishing requirements for the Solid Waste Information and Tracking System (SWITS) as it is converted to a client-server architecture. The purpose is to provide the customer and the performing organizations with the requirements for the SWITS in the new environment. This Software Requirement Specification (SRS) describes the system requirements for the SWITS Project, and follows the PHMC Engineering Requirements, HNF-PRO-1819, and Computer Software Qualify Assurance Requirements, HNF-PRO-309, policies. This SRS includes sections on general description, specific requirements, references, appendices, and index. The SWITS system defined in this document stores information about the solid waste inventory on the Hanford site. Waste is tracked as it is generated, analyzed, shipped, stored, and treated. In addition to inventory reports a number of reports for regulatory agencies are produced.

  2. Spectrum analysis on quality requirements consideration in software design documents.

    Science.gov (United States)

    Kaiya, Haruhiko; Umemura, Masahiro; Ogata, Shinpei; Kaijiri, Kenji

    2013-12-01

    Software quality requirements defined in the requirements analysis stage should be implemented in the final products, such as source codes and system deployment. To guarantee this meta-requirement, quality requirements should be considered in the intermediate stages, such as the design stage or the architectural definition stage. We propose a novel method for checking whether quality requirements are considered in the design stage. In this method, a technique called "spectrum analysis for quality requirements" is applied not only to requirements specifications but also to design documents. The technique enables us to derive the spectrum of a document, and quality requirements considerations in the document are numerically represented in the spectrum. We can thus objectively identify whether the considerations of quality requirements in a requirements document are adapted to its design document. To validate the method, we applied it to commercial software systems with the help of a supporting tool, and we confirmed that the method worked well.

  3. Incubator Display Software Cost Reduction Toolset Software Requirements Specification

    Science.gov (United States)

    Moran, Susanne; Jeffords, Ralph

    2005-01-01

    The Incubator Display Software Requirements Specification was initially developed by Intrinsyx Technologies Corporation (Intrinsyx) under subcontract to Lockheed Martin, Contract Number NAS2-02090, for the National Aeronautics and Space Administration (NASA) Ames Research Center (ARC) Space Station Biological Research Project (SSBRP). The Incubator Display is a User Payload Application (UPA) used to control an Incubator subrack payload for the SSBRP. The Incubator Display functions on-orbit as part of the subrack payload laptop, on the ground as part of the Communication and Data System (CDS) ground control system, and also as part of the crew training environment.

  4. IMAGE information monitoring and applied graphics software environment. Volume 4. Applications description

    International Nuclear Information System (INIS)

    Hallam, J.W.; Ng, K.B.; Upham, G.L.

    1986-09-01

    The EPRI Information Monitoring and Applied Graphics Environment (IMAGE) system is designed for 'fast proto-typing' of advanced concepts for computer-aided plant operations tools. It is a flexible software system which can be used for rapidly creating, dynamically driving and evaluating advanced operator aid displays. The software is written to be both host computer and graphic device independent. This four volume report includes an Executive Overview of the IMAGE package (Volume 1), followed by Software Description (Volume II), User's Guide (Volume III), and Description of Example Applications (Volume IV)

  5. NASA's Software Safety Standard

    Science.gov (United States)

    Ramsay, Christopher M.

    2007-01-01

    NASA relies more and more on software to control, monitor, and verify its safety critical systems, facilities and operations. Since the 1960's there has hardly been a spacecraft launched that does not have a computer on board that will provide command and control services. There have been recent incidents where software has played a role in high-profile mission failures and hazardous incidents. For example, the Mars Orbiter, Mars Polar Lander, the DART (Demonstration of Autonomous Rendezvous Technology), and MER (Mars Exploration Rover) Spirit anomalies were all caused or contributed to by software. The Mission Control Centers for the Shuttle, ISS, and unmanned programs are highly dependant on software for data displays, analysis, and mission planning. Despite this growing dependence on software control and monitoring, there has been little to no consistent application of software safety practices and methodology to NASA's projects with safety critical software. Meanwhile, academia and private industry have been stepping forward with procedures and standards for safety critical systems and software, for example Dr. Nancy Leveson's book Safeware: System Safety and Computers. The NASA Software Safety Standard, originally published in 1997, was widely ignored due to its complexity and poor organization. It also focused on concepts rather than definite procedural requirements organized around a software project lifecycle. Led by NASA Headquarters Office of Safety and Mission Assurance, the NASA Software Safety Standard has recently undergone a significant update. This new standard provides the procedures and guidelines for evaluating a project for safety criticality and then lays out the minimum project lifecycle requirements to assure the software is created, operated, and maintained in the safest possible manner. This update of the standard clearly delineates the minimum set of software safety requirements for a project without detailing the implementation for those

  6. Seven Processes that Enable NASA Software Engineering Technologies

    Science.gov (United States)

    Housch, Helen; Godfrey, Sally

    2011-01-01

    This slide presentation reviews seven processes that NASA uses to ensure that software is developed, acquired and maintained as specified in the NPR 7150.2A requirement. The requirement is to ensure that all software be appraised for the Capability Maturity Model Integration (CMMI). The enumerated processes are: (7) Product Integration, (6) Configuration Management, (5) Verification, (4) Software Assurance, (3) Measurement and Analysis, (2) Requirements Management and (1) Planning & Monitoring. Each of these is described and the group(s) that are responsible is described.

  7. Software Health Management with Bayesian Networks

    Science.gov (United States)

    Mengshoel, Ole; Schumann, JOhann

    2011-01-01

    Most modern aircraft as well as other complex machinery is equipped with diagnostics systems for its major subsystems. During operation, sensors provide important information about the subsystem (e.g., the engine) and that information is used to detect and diagnose faults. Most of these systems focus on the monitoring of a mechanical, hydraulic, or electromechanical subsystem of the vehicle or machinery. Only recently, health management systems that monitor software have been developed. In this paper, we will discuss our approach of using Bayesian networks for Software Health Management (SWHM). We will discuss SWHM requirements, which make advanced reasoning capabilities for the detection and diagnosis important. Then we will present our approach to using Bayesian networks for the construction of health models that dynamically monitor a software system and is capable of detecting and diagnosing faults.

  8. MPS Vax monitor and control software architecture

    International Nuclear Information System (INIS)

    Allison, S.; Spencer, N.; Underwood, K.; VanOlst, D.; Zelanzy, M.

    1993-04-01

    The new Machine Protection System (MPS) now being tested at the SLAC Linear Collider (SLC) includes monitoring and controlling facilities integrated into the existing VAX control system. The actual machine protection is performed by VME micros which control the beam repetition rate on a pulse-by-pulse basis based on measurements from fault detectors. The VAX is used to control and configure the VME micros, configure custom CAMAC modules providing the fault detector inputs, monitor and report faults and system errors, update the SLC database, and interface with the user. The design goals of the VAX software include a database-driven system to allow configuration changes without code changes, use of a standard TCP/IP-based message service for communication, use of existing SLCNET micros for CAMAC configuration, security and verification features to prevent unauthorized access, error and alarm logging and display updates as quickly as possible, and use of touch panels and X-windows displays for the user interface

  9. The Data Quality Monitoring Software for the CMS experiment at the LHC

    CERN Document Server

    AUTHOR|(CDS)2071602

    2016-01-01

    The Data Quality Monitoring (DQM) Software is a central tool in the CMS experiment. Its flexibility allows for integration in several key environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release-Validation, to constantly validate the functionalities and the performance of the reconstruction software; in Monte Carlo productions.Since the end of data taking at a center of mass energy of 8 TeV, the environment in which the DQM lives has undergone fundamental changes. In turn, the DQM system has made significant upgrades in many areas to respond to not only the changes in infrastructure, but also the growing specialized needs of the collaboration with an emphasis on more sophisticated methods for evaluating dataquality, as well as advancing the DQM system to provide quality assessments of various Monte Carlo simulations versus data distributions, monitoring changes in physical effects due to modifications of algorithms or framework, and enabling reg...

  10. Software Assurance in Acquisition: Mitigating Risks to the Enterprise. A Reference Guide for Security-Enhanced Software Acquisition and Outsourcing

    Science.gov (United States)

    2009-02-01

    Monitoring ISO /IEC 12207 2008(E) IEEE 1062 1998 PMBOK 3.0 Initiating Closing 3. Monitoring & Controlling 1. Planning 2. Executing Follow-on...software life cycles [ ISO /IEC 15026]. Software assurance is a key element of national security and homeland security. It is critical because dramatic...they are met. This may also include a plan for testing that SwA requirements are met. The [NDIA] and [ ISO /IEC 15026] provide details on structure and

  11. Questioning the Role of Requirements Engineering in the Causes of Safety-Critical Software Failures

    Science.gov (United States)

    Johnson, C. W.; Holloway, C. M.

    2006-01-01

    Many software failures stem from inadequate requirements engineering. This view has been supported both by detailed accident investigations and by a number of empirical studies; however, such investigations can be misleading. It is often difficult to distinguish between failures in requirements engineering and problems elsewhere in the software development lifecycle. Further pitfalls arise from the assumption that inadequate requirements engineering is a cause of all software related accidents for which the system fails to meet its requirements. This paper identifies some of the problems that have arisen from an undue focus on the role of requirements engineering in the causes of major accidents. The intention is to provoke further debate within the emerging field of forensic software engineering.

  12. Monitoring and reporting software for the coal industry

    Energy Technology Data Exchange (ETDEWEB)

    Okanovic, M. [Advanced Systems Integration Pty Ltd. (Australia)

    2001-08-01

    This paper explains the development and launch of MineSuite software, designed to facilitate report production in coal mines. Advanced Systems Integration (ASI) has developed a system that is generic to all mining operations. Mine personnel can define all processes, KPIs, equipment, delays, reports etc. that are vital in monitoring mining operations. Its capabilities have been realised in opencut, underground and preparation plants throughout Australia. Written in Java, MineSuite is a multi-user, multi-threaded, multi-tasking distributed application. 3 figs.

  13. Active Mirror Predictive and Requirements Verification Software (AMP-ReVS)

    Science.gov (United States)

    Basinger, Scott A.

    2012-01-01

    This software is designed to predict large active mirror performance at various stages in the fabrication lifecycle of the mirror. It was developed for 1-meter class powered mirrors for astronomical purposes, but is extensible to other geometries. The package accepts finite element model (FEM) inputs and laboratory measured data for large optical-quality mirrors with active figure control. It computes phenomenological contributions to the surface figure error using several built-in optimization techniques. These phenomena include stresses induced in the mirror by the manufacturing process and the support structure, the test procedure, high spatial frequency errors introduced by the polishing process, and other process-dependent deleterious effects due to light-weighting of the mirror. Then, depending on the maturity of the mirror, it either predicts the best surface figure error that the mirror will attain, or it verifies that the requirements for the error sources have been met once the best surface figure error has been measured. The unique feature of this software is that it ties together physical phenomenology with wavefront sensing and control techniques and various optimization methods including convex optimization, Kalman filtering, and quadratic programming to both generate predictive models and to do requirements verification. This software combines three distinct disciplines: wavefront control, predictive models based on FEM, and requirements verification using measured data in a robust, reusable code that is applicable to any large optics for ground and space telescopes. The software also includes state-of-the-art wavefront control algorithms that allow closed-loop performance to be computed. It allows for quantitative trade studies to be performed for optical systems engineering, including computing the best surface figure error under various testing and operating conditions. After the mirror manufacturing process and testing have been completed, the

  14. A report on NASA software engineering and Ada training requirements

    Science.gov (United States)

    Legrand, Sue; Freedman, Glenn B.; Svabek, L.

    1987-01-01

    NASA's software engineering and Ada skill base are assessed and information that may result in new models for software engineering, Ada training plans, and curricula are provided. A quantitative assessment which reflects the requirements for software engineering and Ada training across NASA is provided. A recommended implementation plan including a suggested curriculum with associated duration per course and suggested means of delivery is also provided. The distinction between education and training is made. Although it was directed to focus on NASA's need for the latter, the key relationships to software engineering education are also identified. A rationale and strategy for implementing a life cycle education and training program are detailed in support of improved software engineering practices and the transition to Ada.

  15. Metric-based method of software requirements correctness improvement

    Directory of Open Access Journals (Sweden)

    Yaremchuk Svitlana

    2017-01-01

    Full Text Available The work highlights the most important principles of software reliability management (SRM. The SRM concept construes a basis for developing a method of requirements correctness improvement. The method assumes that complicated requirements contain more actual and potential design faults/defects. The method applies a newer metric to evaluate the requirements complexity and double sorting technique evaluating the priority and complexity of a particular requirement. The method enables to improve requirements correctness due to identification of a higher number of defects with restricted resources. Practical application of the proposed method in the course of demands review assured a sensible technical and economic effect.

  16. 7 Processes that Enable NASA Software Engineering Technologies: Value-Added Process Engineering

    Science.gov (United States)

    Housch, Helen; Godfrey, Sally

    2011-01-01

    The presentation reviews Agency process requirements and the purpose, benefits, and experiences or seven software engineering processes. The processes include: product integration, configuration management, verification, software assurance, measurement and analysis, requirements management, and planning and monitoring.

  17. Knowledge-based requirements analysis for automating software development

    Science.gov (United States)

    Markosian, Lawrence Z.

    1988-01-01

    We present a new software development paradigm that automates the derivation of implementations from requirements. In this paradigm, informally-stated requirements are expressed in a domain-specific requirements specification language. This language is machine-understable and requirements expressed in it are captured in a knowledge base. Once the requirements are captured, more detailed specifications and eventually implementations are derived by the system using transformational synthesis. A key characteristic of the process is that the required human intervention is in the form of providing problem- and domain-specific engineering knowledge, not in writing detailed implementations. We describe a prototype system that applies the paradigm in the realm of communication engineering: the prototype automatically generates implementations of buffers following analysis of the requirements on each buffer.

  18. Specification for Visual Requirements of Work-Centered Software Systems

    National Research Council Canada - National Science Library

    Knapp, James R; Chung, Soon M; Schmidt, Vincent A

    2006-01-01

    .... In order to ensure the coherent development and delivery of work-centered software products, WCSS visual requirements must be specified to capture the cognitive aspects of the user interface design...

  19. State of the Art : Integrated Management of Requirements in Model-Based Software Engineering

    OpenAIRE

    Thörn, Christer

    2006-01-01

    This report describes the background and future of research concerning integrated management of requirements in model-based software engineering. The focus is on describing the relevant topics and existing theoretical backgrounds that form the basis for the research. The report describes the fundamental difficulties of requirements engineering for software projects, and proposes that the results and methods of models in software engineering can help leverage those problems. Taking inspiration...

  20. An open software system based on X Windows for process control and equipment monitoring

    International Nuclear Information System (INIS)

    Aimar, A.; Carlier, E.; Mertens, V.

    1992-01-01

    The construction and application of a configurable open software system for process control and equipment monitoring can speed up and simplify the development and maintenance of equipment specific software as compared to individual solutions. The present paper reports the status of such an approach for the distributed control systems of SPS and LEP beam transfer components, based on X Windows and the OSF/Motif tool kit and applying data modeling and software engineering methods. (author)

  1. Groundwater monitoring of hydraulic fracturing in California: Recommendations for permit-required monitoring

    Science.gov (United States)

    Esser, B. K.; Beller, H. R.; Carroll, S.; Cherry, J. A.; Jackson, R. B.; Jordan, P. D.; Madrid, V.; Morris, J.; Parker, B. L.; Stringfellow, W. T.; Varadharajan, C.; Vengosh, A.

    2015-12-01

    California recently passed legislation mandating dedicated groundwater quality monitoring for new well stimulation operations. The authors provided the State with expert advice on the design of such monitoring networks. Factors that must be considered in designing a new and unique groundwater monitoring program include: Program design: The design of a monitoring program is contingent on its purpose, which can range from detection of individual well leakage to demonstration of regional impact. The regulatory goals for permit-required monitoring conducted by operators on a well-by-well basis will differ from the scientific goals of a regional monitoring program conducted by the State. Vulnerability assessment: Identifying factors that increase the probability of transport of fluids from the hydrocarbon target zone to a protected groundwater zone enables the intensity of permit-required monitoring to be tiered by risk and also enables prioritization of regional monitoring of groundwater basins based on vulnerability. Risk factors include well integrity; proximity to existing wellbores and geologic features; wastewater disposal; vertical separation between the hydrocarbon and groundwater zones; and site-specific hydrogeology. Analyte choice: The choice of chemical analytes in a regulatory monitoring program is guided by the goals of detecting impact, assuring public safety, preventing resource degradation, and minimizing cost. Balancing these goals may be best served by tiered approach in which targeted analysis of specific chemical additives is triggered by significant changes in relevant but more easily analyzed constituents. Such an approach requires characterization of baseline conditions, especially in areas with long histories of oil and gas development. Monitoring technology: Monitoring a deep subsurface process or a long wellbore is more challenging than monitoring a surface industrial source. The requirement for monitoring multiple groundwater aquifers across

  2. 40 CFR 429.12 - Monitoring requirements. [Reserved

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 29 2010-07-01 2010-07-01 false Monitoring requirements. [Reserved] 429.12 Section 429.12 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) EFFLUENT... Monitoring requirements. [Reserved] ...

  3. A Software Framework for Remote Patient Monitoring by Using Multi-Agent Systems Support.

    Science.gov (United States)

    Fernandes, Chrystinne Oliveira; Lucena, Carlos José Pereira De

    2017-03-27

    Although there have been significant advances in network, hardware, and software technologies, the health care environment has not taken advantage of these developments to solve many of its inherent problems. Research activities in these 3 areas make it possible to apply advanced technologies to address many of these issues such as real-time monitoring of a large number of patients, particularly where a timely response is critical. The objective of this research was to design and develop innovative technological solutions to offer a more proactive and reliable medical care environment. The short-term and primary goal was to construct IoT4Health, a flexible software framework to generate a range of Internet of things (IoT) applications, containing components such as multi-agent systems that are designed to perform Remote Patient Monitoring (RPM) activities autonomously. An investigation into its full potential to conduct such patient monitoring activities in a more proactive way is an expected future step. A framework methodology was selected to evaluate whether the RPM domain had the potential to generate customized applications that could achieve the stated goal of being responsive and flexible within the RPM domain. As a proof of concept of the software framework's flexibility, 3 applications were developed with different implementations for each framework hot spot to demonstrate potential. Agents4Health was selected to illustrate the instantiation process and IoT4Health's operation. To develop more concrete indicators of the responsiveness of the simulated care environment, an experiment was conducted while Agents4Health was operating, to measure the number of delays incurred in monitoring the tasks performed by agents. IoT4Health's construction can be highlighted as our contribution to the development of eHealth solutions. As a software framework, IoT4Health offers extensibility points for the generation of applications. Applications can extend the framework in

  4. Design Requirements, Epistemic Uncertainty and Solution Development Strategies in Software Design

    DEFF Research Database (Denmark)

    Ball, Linden J.; Onarheim, Balder; Christensen, Bo Thomas

    2010-01-01

    This paper investigates the potential involvement of “epistemic uncertainty” in mediating between complex design requirements and strategic switches in software design strategies. The analysis revealed that the designers produced an initial “first-pass” solution to the given design brief in a bre...... a view of software design as involving a mixed breadth-first and depth-first solution development approach, with strategic switching to depth-first design being triggered by requirement complexity and being mediated by associated feelings of uncertainty....

  5. METHOD FOR SECURITY SPECIFICATION SOFTWARE REQUIREMENTS AS A MEANS FOR IMPLEMENTING A SOFTWARE DEVELOPMENT PROCESS SECURE - MERSEC

    Directory of Open Access Journals (Sweden)

    Castro Mecías, L.T.

    2015-06-01

    Full Text Available Often security incidents that have the object or use the software as a means of causing serious damage and legal, economic consequences, etc. Results of a survey by Kaspersky Lab reflectvulnerabilities in software are the main cause of security incidents in enterprises, the report shows that 85% of them have reported security incidents and vulnerabilities in software are the main reason is further estimated that incidents can cause significant losses estimated from 50,000 to $ 649.000. (1 In this regard academic and industry research focuses on proposals based on reducing vulnerabilities and failures of technology, with a positive influence on how the software is developed. A development process for improved safety practices and should include activities from the initial phases of the software; so that security needs are identified, manage risk and appropriate measures are implemented. This article discusses a method of analysis, acquisition and requirements specification of the software safety analysis on the basis of various proposals and deficiencies identified from participant observation in software development teams. Experiments performed using the proposed yields positive results regarding the reduction of security vulnerabilities and compliance with the safety objectives of the software.

  6. SWATCH Common software for controlling and monitoring the upgraded CMS Level-1 trigger

    CERN Document Server

    Lazaridis, Christos; Bunkowski, Karol; Codispoti, Giuseppe; Dirkx, Glenn; Ghabrous Larrea, Carlos; Lingemann, Joschka; Kreczko, Lukasz; Thea, Alessandro; Williams, Tom

    2017-01-01

    The Large Hadron Collider at CERN restarted in 2015 with a higher centre-of-mass energy of 13 TeV. The instantaneous luminosity is expected to increase significantly in the coming years. An upgraded Level-1 trigger system is being deployed in the CMS experiment in order to maintain the same efficiencies for searches and precision measurements as those achieved in the previous run. This system must be controlled and monitored coherently through software, with high operational efficiency.The legacy system is composed of approximately 4000 data processor boards, of several custom application-specific designs. These boards are organised into several subsystems; each subsystem receives data from different detector systems (calorimeters, barrel/endcap muon detectors), or with differing granularity. These boards have been controlled and monitored by a medium-sized distributed system of over 40 computers and 200 processes. Only a small fraction of the control and monitoring software was common between the different s...

  7. SEJV2 software package for radiation monitoring system of WWER 440 NPP

    International Nuclear Information System (INIS)

    Kapisovsky, V.; Jancik, O.; Kubik, I.; Bena, J.

    1993-01-01

    The main part of the radiation monitoring system at a WWER-440 (213 reactor type) nuclear power plant is the centralized 400-channel monitoring system 'SEJVAL' servicing twin reactor units. The SEJV2 software package is described developed to run on a PC with an IFS2 interface to the SEJVAL radiation monitoring system. It provides enhanced data presentation, record keeping and report generation, thus improving the efficiency of the health physics shift. The system was for the first time implemented at the Jaslovske Bohunice V-2 nuclear power plant with encouraging results. (Z.S.) 3 refs

  8. Basic requirements to the methods of personnel monitoring

    International Nuclear Information System (INIS)

    Keirim-Markus, I.B.

    1981-01-01

    Requirements to methods of personnel monitoring (PMM) depending on irradiation conditions are given. The irradiation conditions determine subjected to monitoring types of irradiation, measurement ranges, periodicity of monitoring, operativeness of obtaining results and required accuracy. The PMM based on the photographic effect of ionizing radiation is the main method of the mass monitoring [ru

  9. Customizable software architectures in the accelerator control system environment

    CERN Document Server

    Mejuev, I; Kadokura, E

    2001-01-01

    Tailoring is further evolution of an application after deployment in order to adapt it to requirements that were not accounted for in the original design. End-user customization has been extensively researched in applied computer science from HCI and software engineering perspectives. Customization allows coping with flexibility requirements, decreasing maintenance and development costs of software products. In general, dynamic or diverse software requirements constitute the need for implementing end-user customization in computer systems. In accelerator physics research the factor of dynamic requirements is especially important, due to frequent software and hardware modifications resulting in correspondingly high upgrade and maintenance costs. We introduce the results of feasibility study on implementing end-user tailorability in the software for accelerator control system, considering the design and implementation of a distributed monitoring application for the 12 GeV KEK Proton Synchrotron as an example. T...

  10. Development of requirements tracking and verification system for the software design of distributed control system

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Chul Hwan; Kim, Jang Yeol; Kim, Jung Tack; Lee, Jang Soo; Ham, Chang Shik [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1999-12-31

    In this paper a prototype of Requirement Tracking and Verification System(RTVS) for a Distributed Control System was implemented and tested. The RTVS is a software design and verification tool. The main functions required by the RTVS are managing, tracking and verification of the software requirements listed in the documentation of the DCS. The analysis of DCS software design procedures and interfaces with documents were performed to define the user of the RTVS, and the design requirements for RTVS were developed. 4 refs., 3 figs. (Author)

  11. Development of requirements tracking and verification system for the software design of distributed control system

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Chul Hwan; Kim, Jang Yeol; Kim, Jung Tack; Lee, Jang Soo; Ham, Chang Shik [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1998-12-31

    In this paper a prototype of Requirement Tracking and Verification System(RTVS) for a Distributed Control System was implemented and tested. The RTVS is a software design and verification tool. The main functions required by the RTVS are managing, tracking and verification of the software requirements listed in the documentation of the DCS. The analysis of DCS software design procedures and interfaces with documents were performed to define the user of the RTVS, and the design requirements for RTVS were developed. 4 refs., 3 figs. (Author)

  12. Software Safety Analysis of Digital Protection System Requirements Using a Qualitative Formal Method

    International Nuclear Information System (INIS)

    Lee, Jang-Soo; Kwon, Kee-Choon; Cha, Sung-Deok

    2004-01-01

    The safety analysis of requirements is a key problem area in the development of software for the digital protection systems of a nuclear power plant. When specifying requirements for software of the digital protection systems and conducting safety analysis, engineers find that requirements are often known only in qualitative terms and that existing fault-tree analysis techniques provide little guidance on formulating and evaluating potential failure modes. A framework for the requirements engineering process is proposed that consists of a qualitative method for requirements specification, called the qualitative formal method (QFM), and a safety analysis method for the requirements based on causality information, called the causal requirements safety analysis (CRSA). CRSA is a technique that qualitatively evaluates causal relationships between software faults and physical hazards. This technique, extending the qualitative formal method process and utilizing information captured in the state trajectory, provides specific guidelines on how to identify failure modes and the relationship among them. The QFM and CRSA processes are described using shutdown system 2 of the Wolsong nuclear power plants as the digital protection system example

  13. Monitoring of computing resource use of active software releases at ATLAS

    Science.gov (United States)

    Limosani, Antonio; ATLAS Collaboration

    2017-10-01

    The LHC is the world’s most powerful particle accelerator, colliding protons at centre of mass energy of 13 TeV. As the energy and frequency of collisions has grown in the search for new physics, so too has demand for computing resources needed for event reconstruction. We will report on the evolution of resource usage in terms of CPU and RAM in key ATLAS offline reconstruction workflows at the TierO at CERN and on the WLCG. Monitoring of workflows is achieved using the ATLAS PerfMon package, which is the standard ATLAS performance monitoring system running inside Athena jobs. Systematic daily monitoring has recently been expanded to include all workflows beginning at Monte Carlo generation through to end-user physics analysis, beyond that of event reconstruction. Moreover, the move to a multiprocessor mode in production jobs has facilitated the use of tools, such as “MemoryMonitor”, to measure the memory shared across processors in jobs. Resource consumption is broken down into software domains and displayed in plots generated using Python visualization libraries and collected into pre-formatted auto-generated Web pages, which allow the ATLAS developer community to track the performance of their algorithms. This information is however preferentially filtered to domain leaders and developers through the use of JIRA and via reports given at ATLAS software meetings. Finally, we take a glimpse of the future by reporting on the expected CPU and RAM usage in benchmark workflows associated with the High Luminosity LHC and anticipate the ways performance monitoring will evolve to understand and benchmark future workflows.

  14. Dynamic visualization techniques for high consequence software

    Energy Technology Data Exchange (ETDEWEB)

    Pollock, G.M.

    1998-02-01

    This report documents a prototype tool developed to investigate the use of visualization and virtual reality technologies for improving software surety confidence. The tool is utilized within the execution phase of the software life cycle. It provides a capability to monitor an executing program against prespecified requirements constraints provided in a program written in the requirements specification language SAGE. The resulting Software Attribute Visual Analysis Tool (SAVAnT) also provides a technique to assess the completeness of a software specification. The prototype tool is described along with the requirements constraint language after a brief literature review is presented. Examples of how the tool can be used are also presented. In conclusion, the most significant advantage of this tool is to provide a first step in evaluating specification completeness, and to provide a more productive method for program comprehension and debugging. The expected payoff is increased software surety confidence, increased program comprehension, and reduced development and debugging time.

  15. Hazard Analysis of Software Requirements Specification for Process Module of FPGA-based Controllers in NPP

    Energy Technology Data Exchange (ETDEWEB)

    Jung; Sejin; Kim, Eui-Sub; Yoo, Junbeom [Konkuk University, Seoul (Korea, Republic of); Keum, Jong Yong; Lee, Jang-Soo [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2016-10-15

    Software in PLC, FPGA which are used to develop I and C system also should be analyzed to hazards and risks before used. NUREG/CR-6430 proposes the method for performing software hazard analysis. It suggests analysis technique for software affected hazards and it reveals that software hazard analysis should be performed with the aspects of software life cycle such as requirements analysis, design, detailed design, implements. It also provides the guide phrases for applying software hazard analysis. HAZOP (Hazard and operability analysis) is one of the analysis technique which is introduced in NUREG/CR-6430 and it is useful technique to use guide phrases. HAZOP is sometimes used to analyze the safety of software. Analysis method of NUREG/CR-6430 had been used in Korea nuclear power plant software for PLC development. Appropriate guide phrases and analysis process are selected to apply efficiently and NUREG/CR-6430 provides applicable methods for software hazard analysis is identified in these researches. We perform software hazard analysis of FPGA software requirements specification with two approaches which are NUREG/CR-6430 and HAZOP with using general GW. We also perform the comparative analysis with them. NUREG/CR-6430 approach has several pros and cons comparing with the HAZOP with general guide words and approach. It is enough applicable to analyze the software requirements specification of FPGA.

  16. Development Of Data Acquisition Software For Centralized Radiation Monitoring System

    International Nuclear Information System (INIS)

    Nolida Yussup; Maslina Mohd Ibrahim; Mohd Fauzi Haris; Syirrazie Che Soh; Harzawardi Hasim; Azraf Azman; Mohd Ashhar Khalid

    2014-01-01

    Nowadays, with the growth of technology, many devices and equipment's can be connected to the network and internet to enable online data acquisition. Centralized radiation monitoring system utilizes a Local Area Network (LAN) as a communication media for data acquisition of the area radiation levels from radiation detectors in Malaysian Nuclear Agency (Nuclear Malaysia). The development of the system involves device configuration, wiring, network and hardware installation, software and web development. This paper describes the software development on the system server that is responsible to acquire and record the area radiation readings from the detectors. Then the recorded readings are called in a web programming to be displayed on a web site. The readings with the time stamp are stored in the system database for query. Besides acquiring the area radiation levels in Nuclear Malaysia centrally, additional features such as data conversion from mR to μSv and line chart display are developed in the software for effective radiation level trend observation and studies. (author)

  17. A Study on the Quantitative Assessment Method of Software Requirement Documents Using Software Engineering Measures and Bayesian Belief Networks

    International Nuclear Information System (INIS)

    Eom, Heung Seop; Kang, Hyun Gook; Park, Ki Hong; Kwon, Kee Choon; Chang, Seung Cheol

    2005-01-01

    One of the major challenges in using the digital systems in a NPP is the reliability estimation of safety critical software embedded in the digital safety systems. Precise quantitative assessment of the reliability of safety critical software is nearly impossible, since many of the aspects to be considered are of qualitative nature and not directly measurable, but they have to be estimated for a practical use. Therefore an expert's judgment plays an important role in estimating the reliability of the software embedded in safety-critical systems in practice, because they can deal with all the diverse evidence relevant to the reliability and can perform an inference based on the evidence. But, in general, the experts' way of combining the diverse evidence and performing an inference is usually informal and qualitative, which is hard to discuss and will eventually lead to a debate about the conclusion. We have been carrying out research on a quantitative assessment of the reliability of safety critical software using Bayesian Belief Networks (BBN). BBN has been proven to be a useful modeling formalism because a user can represent a complex set of events and relationships in a fashion that can easily be interpreted by others. In the previous works we have assessed a software requirement specification of a reactor protection system by using our BBN-based assessment model. The BBN model mainly employed an expert's subjective probabilities as inputs. In the process of assessing the software requirement documents we found out that the BBN model was excessively dependent on experts' subjective judgments in a large part. Therefore, to overcome the weakness of our methodology we employed conventional software engineering measures into the BBN model as shown in this paper. The quantitative relationship between the conventional software measures and the reliability of software were not identified well in the past. Then recently there appeared a few researches on a ranking of

  18. Use of software in the Cuban environmental monitoring

    International Nuclear Information System (INIS)

    Alonso Abad, Dolores; Dominguez Ley, Orlando; Ramos Viltre, Enma O.; Caveda Ramos, Celia; Molina Perez, Daniel; Capote Ferrara, Eduardo; Dominguez Garcia, Adriel; Grinan Torres, Reinaldo

    2008-01-01

    The work group of the National Network of Environmental Radiological Surveillance (NNERS) of the Republic of Cuba offers services which can be divided into two main domains: Services for Environmental Radiological Surveillance and Services for Measuring Radioactivity in Scraps. As part of this work a National Monitoring System has been put into place. Such monitoring system generates a great volume of information which needs to be evaluated, processed, controlled and efficiently stored. To face this challenge in 2002 the National Network started two investigation projects in order to automate some of its services. In this work, we briefly present five software designed to improve the quality of the services offered by the NNERS. We also give a short description of the main results obtained with the use of these tools. The automatization of the measurements for some indicators and the development of technical methods for the interpretation of the results have significantly improved the operational capabilities for the national network. For instance, it has also made possible to carry out studies to characterize radiologically the posts of the network. The improvement of the radiological control of the scrap has diminished the risk of exposure to radiation and the risk of pollution due to the fusion of radioactive sources. The use these software has not only increased technical efficiency of the network, but it has also had a very important social impact. Indeed, the main advantage of such automatization is that it increases the capacity for giving an early response in case of a radiological emergency what redounds in an increased safety for the population and the environment. (author)

  19. The HIE-ISOLDE alignment and monitoring system software and test mock up

    CERN Document Server

    Kautzmann, G; Kadi, Y; Leclercq, Y; Waniorek, S; Williams, L

    2012-01-01

    For the HIE Isolde project a superconducting linac will be built at CERN in the Isolde facility area. The linac will be based on the creation and installation of 2 high- β and 4 low- β cryomodules containing respectively 5 high-β superconducting cavities and 1 superconducting solenoid for the two first ones, 6 low-β superconducting cavities and 2 superconducting solenoids for the four other ones. An alignment and monitoring system of the RF cavities and solenoids placed inside the cryomodules is needed to reach the optimum linac working conditions. The alignment system is based on opto-electronics, optics and precise mechanical instrumentation. The geometrical frame configuration, the data acquisition and the 3D adjustment will be managed using a dedicated software application. In parallel to the software development, an alignment system test mock-up has been built for software validation and dimensional tests. This paper will present the software concept and the development status, and then will describe...

  20. Designing for Change: Minimizing the Impact of Changing Requirements in the Later Stages of a Spaceflight Software Project

    Science.gov (United States)

    Allen, B. Danette

    1998-01-01

    In the traditional 'waterfall' model of the software project life cycle, the Requirements Phase ends and flows into the Design Phase, which ends and flows into the Development Phase. Unfortunately, the process rarely, if ever, works so smoothly in practice. Instead, software developers often receive new requirements, or modifications to the original requirements, well after the earlier project phases have been completed. In particular, projects with shorter than ideal schedules are highly susceptible to frequent requirements changes, as the software requirements analysis phase is often forced to begin before the overall system requirements and top-level design are complete. This results in later modifications to the software requirements, even though the software design and development phases may be complete. Requirements changes received in the later stages of a software project inevitably lead to modification of existing developed software. Presented here is a series of software design techniques that can greatly reduce the impact of last-minute requirements changes. These techniques were successfully used to add built-in flexibility to two complex software systems in which the requirements were expected to (and did) change frequently. These large, real-time systems were developed at NASA Langley Research Center (LaRC) to test and control the Lidar In-Space Technology Experiment (LITE) instrument which flew aboard the space shuttle Discovery as the primary payload on the STS-64 mission.

  1. Comparison of monitor units calculated by radiotherapy treatment planning system and an independent monitor unit verification software.

    Science.gov (United States)

    Sellakumar, P; Arun, C; Sanjay, S S; Ramesh, S B

    2011-01-01

    In radiation therapy, the monitor units (MU) needed to deliver a treatment plan are calculated by treatment planning systems (TPS). The essential part of quality assurance is to verify the MU with independent monitor unit calculation to correct any potential errors prior to the start of treatment. In this study, we have compared the MU calculated by TPS and by independent MU verification software. The MU verification software was commissioned and tested for the data integrity to ensure that the correct beam data was considered for MU calculations. The accuracy of the calculations was tested by creating a series of test plans and comparing them with ion chamber measurements. The results show that there is good agreement between the two. The MU difference (MUdiff) between the monitor unit calculations of TPS and independent MU verification system was calculated for 623 fields from 245 patients and was analyzed by treatment site for head & neck, thorax, breast, abdomen and pelvis. The mean MUdiff of -0.838% with a standard deviation of 3.04% was observed for all 623 fields. The site specific standard deviation of MUdiff was as follows: abdomen and pelvis (<1.75%), head & neck (2.5%), thorax (2.32%) and breast (6.01%). The disparities were analyzed and different correction methods were used to reduce the disparity. © 2010 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  2. 40 CFR 141.74 - Analytical and monitoring requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 22 2010-07-01 2010-07-01 false Analytical and monitoring requirements... Analytical and monitoring requirements. (a) Analytical requirements. Only the analytical method(s) specified... as set forth in the article “National Field Evaluation of a Defined Substrate Method for the...

  3. ATLAS Offline Software Performance Monitoring and Optimization

    CERN Document Server

    Chauhan, N; Kittelmann, T; Langenberg, R; Mandrysch , R; Salzburger, A; Seuster, R; Ritsch, E; Stewart, G; van Eldik, N; Vitillo, R

    2014-01-01

    In a complex multi-developer, multi-package software environment, such as the ATLAS offline Athena framework, tracking the performance of the code can be a non-trivial task in itself. In this paper we describe improvements in the instrumentation of ATLAS offline software that have given considerable insight into the performance of the code and helped to guide optimisation. Code can be instrumented firstly using the PAPI tool, which is a programing interface for accessing hardware performance counters. PAPI events can count floating point operations, cycles and instructions and cache accesses. Triggering PAPI to start/stop counting for each algorithm and processed event gives a good understanding of the whole algorithm level performance of ATLAS code. Further data can be obtained using pin, a dynamic binary instrumentation tool. Pintools can be used to obtain similar statistics as PAPI, but advantageously without requiring recompilation of the code. Fine grained routine and instruction level instrumentation is...

  4. ATLAS Offline Software Performance Monitoring and Optimization

    CERN Document Server

    Chauhan, N; The ATLAS collaboration; Kittelmann, T; Langenberg, R; Mandrysch , R; Salzburger, A; Seuster, R; Ritsch, E; Stewart, G; van Eldik, N; Vitillo, R

    2013-01-01

    In a complex multi-developer, multi-package software environment, such as the ATLAS offline Athena framework, tracking the performance of the code can be a non-trivial task in itself. In this paper we describe improvements in the instrumentation of ATLAS offline software that have given considerable insight into the performance of the code and helped to guide optimisation. Code can be instrumented firstly using the PAPI tool, which is a programing interface for accessing hardware performance counters. PAPI events can count floating point operations, cycles and instructions and cache accesses. Triggering PAPI to start/stop counting for each algorithm and processed event gives a good understanding of the whole algorithm level performance of ATLAS code. Further data can be obtained using pin, a dynamic binary instrumentation tool. Pintools can be used to obtain similar statistics as PAPI, but advantageously without requiring recompilation of the code. Fine grained routine and instruction level instrumentation is...

  5. The Database and Data Analysis Software of Radiation Monitoring System

    International Nuclear Information System (INIS)

    Wang Weizhen; Li Jianmin; Wang Xiaobing; Hua Zhengdong; Xu Xunjiang

    2009-01-01

    Shanghai Synchrotron Radiation Facility (SSRF for short) is a third-generation light source building in China, including a 150MeV injector, 3.5GeV booster, 3.5GeV storage ring and an amount of beam line stations. The data is fetched by the monitoring computer from collecting modules in the front end, and saved in the MySQL database in the managing computer. The data analysis software is coded with Python, a script language, to inquire, summarize and plot the data of a certain monitoring channel during a certain period and export to an external file. In addition, the warning event can be inquired separately. The website for historical and real-time data inquiry and plotting is coded with PHP. (authors)

  6. Monitoring of Computing Resource Use of Active Software Releases in ATLAS

    CERN Document Server

    Limosani, Antonio; The ATLAS collaboration

    2016-01-01

    The LHC is the world's most powerful particle accelerator, colliding protons at centre of mass energy of 13 TeV. As the energy and frequency of collisions has grown in the search for new physics, so too has demand for computing resources needed for event reconstruction. We will report on the evolution of resource usage in terms of CPU and RAM in key ATLAS offline reconstruction workflows at the Tier0 at CERN and on the WLCG. Monitoring of workflows is achieved using the ATLAS PerfMon package, which is the standard ATLAS performance monitoring system running inside Athena jobs. Systematic daily monitoring has recently been expanded to include all workflows beginning at Monte Carlo generation through to end user physics analysis, beyond that of event reconstruction. Moreover, the move to a multiprocessor mode in production jobs has facilitated the use of tools, such as "MemoryMonitor", to measure the memory shared across processors in jobs. Resource consumption is broken down into software domains and displayed...

  7. Upgrading of data acquisition software for centralized radiation monitoring system in Malaysian Nuclear Agency

    Energy Technology Data Exchange (ETDEWEB)

    Yussup, F., E-mail: nolida@nm.gov.my; Ibrahim, M. M., E-mail: maslina-i@nm.gov.my; Soh, S. C.; Hasim, H. [Instrumentation and Automation Centre, Technical Support Division, Malaysian Nuclear Agency, Bangi, 43000 Kajang. Selangor (Malaysia); Haris, M. F. [Information Technology Centre, Technical Support Division, Malaysian Nuclear Agency, Bangi, 43000 Kajang. Selangor (Malaysia); Azman, A. [Prototype and Development Centre, Technical Support Division, Malaysian Nuclear Agency, Bangi, 43000 Kajang. Selangor (Malaysia); Razalim, F. A. A.; Yapp, R. [Health Physics Group, Radiation Safety and Health Division, Malaysian Nuclear Agency, Bangi, 43000 Kajang. Selangor (Malaysia); Ramli, A. A. M. [Technical Support Division, Malaysian Nuclear Agency, Bangi, 43000 Kajang. Selangor (Malaysia)

    2016-01-22

    With the growth of technology, many devices and equipments can be connected to the network and internet to enable online data acquisition for real-time data monitoring and control from monitoring devices located at remote sites. Centralized radiation monitoring system (CRMS) is a system that enables area radiation level at various locations in Malaysian Nuclear Agency (Nuklear Malaysia) to be monitored centrally by using a web browser. The Local Area Network (LAN) in Nuclear Malaysia is utilized in CRMS as a communication media for data acquisition of the area radiation levels from radiation detectors. The development of the system involves device configuration, wiring, network and hardware installation, software and web development. This paper describes the software upgrading on the system server that is responsible to acquire and record the area radiation readings from the detectors. The recorded readings are called in a web programming to be displayed on a website. Besides the main feature which is acquiring the area radiation levels in Nuclear Malaysia centrally, the upgrading involves new features such as uniform time interval for data recording and exporting, warning system and dose triggering.

  8. Upgrading of data acquisition software for centralized radiation monitoring system in Malaysian Nuclear Agency

    International Nuclear Information System (INIS)

    Yussup, F.; Ibrahim, M. M.; Soh, S. C.; Hasim, H.; Haris, M. F.; Azman, A.; Razalim, F. A. A.; Yapp, R.; Ramli, A. A. M.

    2016-01-01

    With the growth of technology, many devices and equipments can be connected to the network and internet to enable online data acquisition for real-time data monitoring and control from monitoring devices located at remote sites. Centralized radiation monitoring system (CRMS) is a system that enables area radiation level at various locations in Malaysian Nuclear Agency (Nuklear Malaysia) to be monitored centrally by using a web browser. The Local Area Network (LAN) in Nuclear Malaysia is utilized in CRMS as a communication media for data acquisition of the area radiation levels from radiation detectors. The development of the system involves device configuration, wiring, network and hardware installation, software and web development. This paper describes the software upgrading on the system server that is responsible to acquire and record the area radiation readings from the detectors. The recorded readings are called in a web programming to be displayed on a website. Besides the main feature which is acquiring the area radiation levels in Nuclear Malaysia centrally, the upgrading involves new features such as uniform time interval for data recording and exporting, warning system and dose triggering

  9. Clustering and Flow Conservation Monitoring Tool for Software Defined Networks

    Directory of Open Access Journals (Sweden)

    Jesús Antonio Puente Fernández

    2018-04-01

    Full Text Available Prediction systems present some challenges on two fronts: the relation between video quality and observed session features and on the other hand, dynamics changes on the video quality. Software Defined Networks (SDN is a new concept of network architecture that provides the separation of control plane (controller and data plane (switches in network devices. Due to the existence of the southbound interface, it is possible to deploy monitoring tools to obtain the network status and retrieve a statistics collection. Therefore, achieving the most accurate statistics depends on a strategy of monitoring and information requests of network devices. In this paper, we propose an enhanced algorithm for requesting statistics to measure the traffic flow in SDN networks. Such an algorithm is based on grouping network switches in clusters focusing on their number of ports to apply different monitoring techniques. Such grouping occurs by avoiding monitoring queries in network switches with common characteristics and then, by omitting redundant information. In this way, the present proposal decreases the number of monitoring queries to switches, improving the network traffic and preventing the switching overload. We have tested our optimization in a video streaming simulation using different types of videos. The experiments and comparison with traditional monitoring techniques demonstrate the feasibility of our proposal maintaining similar values decreasing the number of queries to the switches.

  10. Clustering and Flow Conservation Monitoring Tool for Software Defined Networks.

    Science.gov (United States)

    Puente Fernández, Jesús Antonio; García Villalba, Luis Javier; Kim, Tai-Hoon

    2018-04-03

    Prediction systems present some challenges on two fronts: the relation between video quality and observed session features and on the other hand, dynamics changes on the video quality. Software Defined Networks (SDN) is a new concept of network architecture that provides the separation of control plane (controller) and data plane (switches) in network devices. Due to the existence of the southbound interface, it is possible to deploy monitoring tools to obtain the network status and retrieve a statistics collection. Therefore, achieving the most accurate statistics depends on a strategy of monitoring and information requests of network devices. In this paper, we propose an enhanced algorithm for requesting statistics to measure the traffic flow in SDN networks. Such an algorithm is based on grouping network switches in clusters focusing on their number of ports to apply different monitoring techniques. Such grouping occurs by avoiding monitoring queries in network switches with common characteristics and then, by omitting redundant information. In this way, the present proposal decreases the number of monitoring queries to switches, improving the network traffic and preventing the switching overload. We have tested our optimization in a video streaming simulation using different types of videos. The experiments and comparison with traditional monitoring techniques demonstrate the feasibility of our proposal maintaining similar values decreasing the number of queries to the switches.

  11. Software requirements definition Shipping Cask Analysis System (SCANS)

    International Nuclear Information System (INIS)

    Johnson, G.L.; Serbin, R.

    1985-01-01

    The US Nuclear Regulatory Commission (NRC) staff reviews the technical adequacy of applications for certification of designs of shipping casks for spent nuclear fuel. In order to confirm an acceptable design, the NRC staff may perform independent calculations. The current NRC procedure for confirming cask design analyses is laborious and tedious. Most of the work is currently done by hand or through the use of a remote computer network. The time required to certify a cask can be long. The review process may vary somewhat with the engineer doing the reviewing. Similarly, the documentation on the results of the review can also vary with the reviewer. To increase the efficiency of this certification process, LLNL was requested to design and write an integrated set of user-oriented, interactive computer programs for a personal microcomputer. The system is known as the NRC Shipping Cask Analysis System (SCANS). The computer codes and the software system supporting these codes are being developed and maintained for the NRC by LLNL. The objective of this system is generally to lessen the time and effort needed to review an application. Additionally, an objective of the system is to assure standardized methods and documentation of the confirmatory analyses used in the review of these cask designs. A software system should be designed based on NRC-defined requirements contained in a requirements document. The requirements document is a statement of a project's wants and needs as the users and implementers jointly understand them. The requirements document states the desired end products (i.e. WHAT's) of the project, not HOW the project provides them. This document describes the wants and needs for the SCANS system. 1 fig., 3 tabs

  12. The development of test software for the inadequate core cooling monitoring system

    International Nuclear Information System (INIS)

    Lee, Soon Sung.

    1996-06-01

    The test software including the ICCMS simulator which is necessary for dynamic test for the ICCMS software in PWR is developed. The developed dynamic test software consists of the module test simulator, the integration test simulator, and the test result analyser. The simulator was programmed by C language according to the same algorithm requirements for the FORTRAN version ICCMS software, and also for the Factory Acceptance Test (FAT). And the simulator can be used as training tool for the reactor operator and system development tool for the performance improvement. (author). 4 tabs., 8 figs., 11 refs

  13. Monitoring of performance management using Quality Assurance Indicators and ISO requirement

    Directory of Open Access Journals (Sweden)

    Dargahi H

    2007-06-01

    Full Text Available Background: Quality assurance is a prevention-oriented system that can be used to improve the quality of care, increase productivity and monitor the performance management in clinical laboratories. ISO 9001: 2000 requirements are a collection of management and technical systems designed to implement quality assurance and monitor performance management in organizations. Methods: A checklist was prepared to monitor the preanalytical, analytical and postanalytical stages of laboratory performance management in 16 areas and all laboratory activities in 14 of the clinical laboratories of the Tehran University of Medical Sciences (TUMS hospitals. Collected data were stored and statistically analyzed using SPSS software. Results: The best performance, in which 77.73% of quality assurance indicators were observed, was found in Sina Hospital. However, only 57.56% of these indicators were fulfilled at Farabi Hospital, with the lowest-level performance among the clinical laboratories of TUMS hospitals. The highest level of compliance with quality assurance indicators was in the hematology departments and for facility demands in management areas. Overall, quality assurance indicators were appropriately followed in only 7% of the clinical laboratories. Conclusion: The average quality assurance observation rate in the clinical laboratories studied was 67.22%, which is insufficient and must be remedied with stricter enforcement of the ISO 9001: 2000 regulations.

  14. A signal processing application for evaluating self-monitoring blood glucose strategies in a software agent model.

    Science.gov (United States)

    Wang, Zhanle; Paranjape, Raman

    2015-07-01

    We propose the signal processing technique of calculating a cross-correlation function and an average deviation between the continuous blood glucose and the interpolation of limited blood glucose samples to evaluate blood glucose monitoring frequency in a self-aware patient software agent model. The diabetic patient software agent model [1] is a 24-h circadian, self-aware, stochastic model of a diabetic patient's blood glucose levels in a software agent environment. The purpose of this work is to apply a signal processing technique to assist patients and physicians in understanding the extent of a patient's illness using a limited number of blood glucose samples. A second purpose of this work is to determine an appropriate blood glucose monitoring frequency in order to have a minimum number of samples taken that still provide a good understanding of the patient's blood glucose levels. For society in general, the monitoring cost of diabetes is an extremely important issue, and these costs can vary tremendously depending on monitoring approaches and monitoring frequencies. Due to the cost and discomfort associated with blood glucose monitoring, today, patients expect monitoring frequencies specific to their health profile. The proposed method quantitatively assesses various monitoring protocols (from 6 times per day to 1 time per week) in nine predefined categories of patient agents in terms of risk factors of health status and age. Simulation results show that sampling 6 times per day is excessive, and not necessary for understanding the dynamics of the continuous signal in the experiments. In addition, patient agents in certain conditions only need to sample their blood glucose 1 time per week to have a good understanding of the characteristics of their blood glucose. Finally, an evaluation scenario is developed to visualize this concept, in which appropriate monitoring frequencies are shown based on the particular conditions of patient agents. This base line can

  15. Analysis of free geo-server software usability from the viewpoint of INSPIRE requirementsAnalysis of free geo-server software usability from the viewpoint of INSPIRE requirements

    Directory of Open Access Journals (Sweden)

    Tomasz  Grasza

    2014-06-01

    Full Text Available The paper presents selected server platforms based on free and open source license, coherent with the standards of the Open Geospatial Consortium. The presented programs are evaluated in the context of the INSPIRE Directive. The first part describes the requirements of the Directive, and afterwards presented are the pros and cons of each platform, to meet these demands. This article provides an answer to the question whether the use of free software can provide interoperable network services in accordance with the requirements of the INSPIRE Directive, on the occasion of presenting the application examples and practical tips on the use of particular programs.[b]Keywords[/b]: GIS, INSPIRE, free software, OGC, geoportal, network services, GeoServer, deegree, GeoNetwork

  16. Reliability design of the continuous monitoring system software for an position radiation

    International Nuclear Information System (INIS)

    Kang Yuebing; Li Tiantuo; Di Yuming; Zhang Yanhong

    2004-01-01

    The reliability and stabilization is an important technical target for a continuous monitoring system. After analyzing the position's environment and the system's structure, we put forward some methods of the software's reliability design and put these into the application. The practice shows that it is important to improve the system's stability and reliability. (authors)

  17. Effective organizational solutions for implementation of DBMS software packages

    Science.gov (United States)

    Jones, D.

    1984-01-01

    The space telescope management information system development effort is a guideline for discussing effective organizational solutions used in implementing DBMS software. Focus is on the importance of strategic planning. The value of constructing an information system architecture to conform to the organization's managerial needs, the need for a senior decision maker, dealing with shifting user requirements, and the establishment of a reliable working relationship with the DBMS vendor are examined. Requirements for a schedule to demonstrate progress against a defined timeline and the importance of continued monitoring for production software control, production data control, and software enhancements are also discussed.

  18. Software quality assurance and software safety in the Biomed Control System

    International Nuclear Information System (INIS)

    Singh, R.P.; Chu, W.T.; Ludewigt, B.A.; Marks, K.M.; Nyman, M.A.; Renner, T.R.; Stradtner, R.

    1989-01-01

    The Biomed Control System is a hardware/software system used for the delivery, measurement and monitoring of heavy-ion beams in the patient treatment and biology experiment rooms in the Bevalac at the Lawrence Berkeley Laboratory (LBL). This paper describes some aspects of this system including historical background philosophy, configuration management, hardware features that facilitate software testing, software testing procedures, the release of new software quality assurance, safety and operator monitoring. 3 refs

  19. Dependencies among Architectural Views Got from Software Requirements Based on a Formal Model

    Directory of Open Access Journals (Sweden)

    Osis Janis

    2014-12-01

    Full Text Available A system architect has software requirements and some unspecified knowledge about a problem domain (e.g., an enterprise as source information for assessment and evaluation of possible solutions and getting the target point, a preliminary software design. The solving factor is architect’s experience and expertise in the problem domain (“AS-IS”. A proposed approach is dedicated to assist a system architect in making an appropriate decision on the solution (“TO-BE”. It is based on a formal mathematical model, Topological Functioning Model (TFM. Compliant TFMs can be transformed into software architectural views. The paper demonstrates and discusses tracing dependency links from the requirements to and between the architectural views.

  20. Self-assembled software and method of overriding software execution

    Science.gov (United States)

    Bouchard, Ann M.; Osbourn, Gordon C.

    2013-01-08

    A computer-implemented software self-assembled system and method for providing an external override and monitoring capability to dynamically self-assembling software containing machines that self-assemble execution sequences and data structures. The method provides an external override machine that can be introduced into a system of self-assembling machines while the machines are executing such that the functionality of the executing software can be changed or paused without stopping the code execution and modifying the existing code. Additionally, a monitoring machine can be introduced without stopping code execution that can monitor specified code execution functions by designated machines and communicate the status to an output device.

  1. Advanced software tools for digital loose part monitoring systems

    International Nuclear Information System (INIS)

    Ding, Y.

    1996-01-01

    The paper describes two software modules as analysis tools for digital loose part monitoring systems. The first module is called acoustic module which utilizes the multi-media features of modern personal computers to replay the digital stored short-time bursts with sufficient length and in good quality. This is possible due to the so-called puzzle technique developed at ISTec. The second module is called classification module which calculates advanced burst parameters and classifies the acoustic events in pre-defined classes with the help of an artificial multi-layer perception neural network trained with the back propagation algorithm. (author). 7 refs, 7 figs

  2. Advanced software tools for digital loose part monitoring systems

    Energy Technology Data Exchange (ETDEWEB)

    Ding, Y [Institute for Safety Technology (ISTec) GmbH, Garching (Germany)

    1997-12-31

    The paper describes two software modules as analysis tools for digital loose part monitoring systems. The first module is called acoustic module which utilizes the multi-media features of modern personal computers to replay the digital stored short-time bursts with sufficient length and in good quality. This is possible due to the so-called puzzle technique developed at ISTec. The second module is called classification module which calculates advanced burst parameters and classifies the acoustic events in pre-defined classes with the help of an artificial multi-layer perception neural network trained with the back propagation algorithm. (author). 7 refs, 7 figs.

  3. Towards an Early Software Effort Estimation Based on Functional and Non-Functional Requirements

    NARCIS (Netherlands)

    Kassab, M.; Daneva, Maia; Ormanjieva, Olga; Abran, A.; Braungarten, R.; Dumke, R.; Cuadrado-Gallego, J.; Brunekreef, J.

    2009-01-01

    The increased awareness of the non-functional requirements as a key to software project and product success makes explicit the need to include them in any software project effort estimation activity. However, the existing approaches to defining size-based effort relationships still pay insufficient

  4. Towards an Early Software Effort Estimation Based on Functional and Non-Functional Requirements

    Science.gov (United States)

    Kassab, Mohamed; Daneva, Maya; Ormandjieva, Olga

    The increased awareness of the non-functional requirements as a key to software project and product success makes explicit the need to include them in any software project effort estimation activity. However, the existing approaches to defining size-based effort relationships still pay insufficient attention to this need. This paper presents a flexible, yet systematic approach to the early requirements-based effort estimation, based on Non-Functional Requirements ontology. It complementarily uses one standard functional size measurement model and a linear regression technique. We report on a case study which illustrates the application of our solution approach in context and also helps evaluate our experiences in using it.

  5. Antenna Controller Replacement Software

    Science.gov (United States)

    Chao, Roger Y.; Morgan, Scott C.; Strain, Martha M.; Rockwell, Stephen T.; Shimizu, Kenneth J.; Tehrani, Barzia J.; Kwok, Jaclyn H.; Tuazon-Wong, Michelle; Valtier, Henry; Nalbandi, Reza; hide

    2010-01-01

    The Antenna Controller Replacement (ACR) software accurately points and monitors the Deep Space Network (DSN) 70-m and 34-m high-efficiency (HEF) ground-based antennas that are used to track primarily spacecraft and, periodically, celestial targets. To track a spacecraft, or other targets, the antenna must be accurately pointed at the spacecraft, which can be very far away with very weak signals. ACR s conical scanning capability collects the signal in a circular pattern around the target, calculates the location of the strongest signal, and adjusts the antenna pointing to point directly at the spacecraft. A real-time, closed-loop servo control algorithm performed every 0.02 second allows accurate positioning of the antenna in order to track these distant spacecraft. Additionally, this advanced servo control algorithm provides better antenna pointing performance in windy conditions. The ACR software provides high-level commands that provide a very easy user interface for the DSN operator. The operator only needs to enter two commands to start the antenna and subreflector, and Master Equatorial tracking. The most accurate antenna pointing is accomplished by aligning the antenna to the Master Equatorial, which because of its small size and sheltered location, has the most stable pointing. The antenna has hundreds of digital and analog monitor points. The ACR software provides compact displays to summarize the status of the antenna, subreflector, and the Master Equatorial. The ACR software has two major functions. First, it performs all of the steps required to accurately point the antenna (and subreflector and Master Equatorial) at the spacecraft (or celestial target). This involves controlling the antenna/ subreflector/Master-Equatorial hardware, initiating and monitoring the correct sequence of operations, calculating the position of the spacecraft relative to the antenna, executing the real-time servo control algorithm to maintain the correct position, and

  6. Requirements for guidelines systems: implementation challenges and lessons from existing software-engineering efforts.

    Science.gov (United States)

    Shah, Hemant; Allard, Raymond D; Enberg, Robert; Krishnan, Ganesh; Williams, Patricia; Nadkarni, Prakash M

    2012-03-09

    A large body of work in the clinical guidelines field has identified requirements for guideline systems, but there are formidable challenges in translating such requirements into production-quality systems that can be used in routine patient care. Detailed analysis of requirements from an implementation perspective can be useful in helping define sub-requirements to the point where they are implementable. Further, additional requirements emerge as a result of such analysis. During such an analysis, study of examples of existing, software-engineering efforts in non-biomedical fields can provide useful signposts to the implementer of a clinical guideline system. In addition to requirements described by guideline-system authors, comparative reviews of such systems, and publications discussing information needs for guideline systems and clinical decision support systems in general, we have incorporated additional requirements related to production-system robustness and functionality from publications in the business workflow domain, in addition to drawing on our own experience in the development of the Proteus guideline system (http://proteme.org). The sub-requirements are discussed by conveniently grouping them into the categories used by the review of Isern and Moreno 2008. We cite previous work under each category and then provide sub-requirements under each category, and provide example of similar work in software-engineering efforts that have addressed a similar problem in a non-biomedical context. When analyzing requirements from the implementation viewpoint, knowledge of successes and failures in related software-engineering efforts can guide implementers in the choice of effective design and development strategies.

  7. Common software for controlling and monitoring the upgraded CMS Level-1 trigger

    CERN Document Server

    Codispoti, Giuseppe

    2017-01-01

    The Large Hadron Collider restarted in 2015 with a higher centre-of-mass energy of 13 TeV. The instantaneous luminosity is expected to increase significantly in the coming years. An upgraded Level-1 trigger system was deployed in the CMS experiment in order to maintain the same efficiencies for searches and precision measurements as those achieved in 2012. This system must be controlled and monitored coherently through software, with high operational efficiency.The legacy system was composed of a large number of custom data processor boards; correspondingly, only a small fraction of the software was common between the different subsystems. The upgraded system is composed of a set of general purpose boards, that follow the MicroTCA specification, and transmit data over optical links, resulting in a more homogeneous system. The associated software is based on generic components corresponding to the firmware blocks that are shared across different cards, regardless of the role that the card plays in the system. ...

  8. Knowledge Base for an Intelligent System in order to Identify Security Requirements for Government Agencies Software Projects

    Directory of Open Access Journals (Sweden)

    Adán Beltrán G.

    2016-01-01

    Full Text Available It has been evidenced that one of the most common causes in the failure of software security is the lack of identification and specification of requirements for information security, it is an activity with an insufficient importance in the software development or software acquisition We propose the knowledge base of CIBERREQ. CIBERREQ is an intelligent knowledge-based system used for the identification and specification of security requirements in the software development cycle or in the software acquisition. CIBERREQ receives functional software requirements written in natural language and produces non-functional security requirements through a semi-automatic process of risk management. The knowledge base built is formed by an ontology developed collaboratively by experts in information security. In this process has been identified six types of assets: electronic data, physical data, hardware, software, person and service; as well as six types of risk: competitive disadvantage, loss of credibility, economic risks, strategic risks, operational risks and legal sanctions. In addition there are defined 95 vulnerabilities, 24 threats, 230 controls, and 515 associations between concepts. Additionally, automatic expansion was used with Wikipedia for the asset types Software and Hardware, obtaining 7125 and 5894 software and hardware subtypes respectively, achieving thereby an improvement of 10% in the identification of the information assets candidates, one of the most important phases of the proposed system.

  9. Development to requirements for a procedures software tool

    International Nuclear Information System (INIS)

    Yasutake, J.Y.; Hachiro Isoda

    1993-01-01

    In 1989, the Electric Power Research Institute (EPRI) and the Central Research Institute of the Electric Power Industry (CRIEPI) in Japan initiated a joint research program to investigate various interventions to reduce personnel errors and inefficiencies in the maintenance of nuclear power plants. This program, consisting of several interrelated projects, was initiated because of the mutual recognition of the importance of the human element in the efficient and safe operation of utilities and the continuing need to enhance personnel performance to sustain plant safety and availability. This paper summarizes one of the projects, jointly funded by EPRI and CRIEPI, to analyze the requirements for, and prepare a functional description of, a procedures software tool (PST). The primary objective of this project was to develop a description of the features and functions of a software tool that would help procedure writers to improve the quality of maintenance and testing procedures, thereby enhancing the performance of both procedure writers and maintenance personnel

  10. Autonomous Real Time Requirements Tracing

    Science.gov (United States)

    Plattsmier, George; Stetson, Howard

    2014-01-01

    One of the more challenging aspects of software development is the ability to verify and validate the functional software requirements dictated by the Software Requirements Specification (SRS) and the Software Detail Design (SDD). Insuring the software has achieved the intended requirements is the responsibility of the Software Quality team and the Software Test team. The utilization of Timeliner-TLX(sup TM) Auto- Procedures for relocating ground operations positions to ISS automated on-board operations has begun the transition that would be required for manned deep space missions with minimal crew requirements. This transition also moves the auto-procedures from the procedure realm into the flight software arena and as such the operational requirements and testing will be more structured and rigorous. The autoprocedures would be required to meet NASA software standards as specified in the Software Safety Standard (NASASTD- 8719), the Software Engineering Requirements (NPR 7150), the Software Assurance Standard (NASA-STD-8739) and also the Human Rating Requirements (NPR-8705). The Autonomous Fluid Transfer System (AFTS) test-bed utilizes the Timeliner-TLX(sup TM) Language for development of autonomous command and control software. The Timeliner-TLX(sup TM) system has the unique feature of providing the current line of the statement in execution during real-time execution of the software. The feature of execution line number internal reporting unlocks the capability of monitoring the execution autonomously by use of a companion Timeliner-TLX(sup TM) sequence as the line number reporting is embedded inside the Timeliner-TLX(sup TM) execution engine. This negates I/O processing of this type data as the line number status of executing sequences is built-in as a function reference. This paper will outline the design and capabilities of the AFTS Autonomous Requirements Tracker, which traces and logs SRS requirements as they are being met during real-time execution of the

  11. Tailorable software architectures in the accelerator control system environment

    International Nuclear Information System (INIS)

    Mejuev, Igor; Kumagai, Akira; Kadokura, Eiichi

    2001-01-01

    Tailoring is further evolution of an application after deployment in order to adapt it to requirements that were not accounted for in the original design. End-user tailorability has been extensively researched in applied computer science from HCl and software engineering perspectives. Tailorability allows coping with flexibility requirements, decreasing maintenance and development costs of software products. In general, dynamic or diverse software requirements constitute the need for implementing end-user tailorability in computer systems. In accelerator physics research the factor of dynamic requirements is especially important, due to frequent software and hardware modifications resulting in correspondingly high upgrade and maintenance costs. In this work we introduce the results of feasibility study on implementing end-user tailorability in the software for accelerator control system, considering the design and implementation of distributed monitoring application for 12 GeV KEK Proton Synchrotron as an example. The software prototypes used in this work are based on a generic tailoring platform (VEDICI), which allows decoupling of tailoring interfaces and runtime components. While representing a reusable application-independent framework, VEDICI can be potentially applied for tailoring of arbitrary compositional Web-based applications

  12. Independent Verification and Validation Of SAPHIRE 8 Software Requirements Project Number: N6423 U.S. Nuclear Regulatory Commission

    Energy Technology Data Exchange (ETDEWEB)

    Kent Norris

    2009-09-01

    The purpose of the Independent Verification and Validation (IV&V) role in the evaluation of the SAPHIRE requirements definition is to assess the activities that results in the specification, documentation, and review of the requirements that the software product must satisfy, including functionality, performance, design constraints, attributes and external interfaces. The IV&V team began this endeavor after the software engineering and software development of SAPHIRE had already been in production. IV&V reviewed the requirements specified in the NRC Form 189s to verify these requirements were included in SAPHIRE’s Software Verification and Validation Plan (SVVP).

  13. 40 CFR 410.02 - Monitoring requirements. [Reserved

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 28 2010-07-01 2010-07-01 true Monitoring requirements. [Reserved] 410.02 Section 410.02 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) EFFLUENT... requirements. [Reserved] ...

  14. Attributes Of Quality Scenarios/Scenario Sets Used In Software Requirements Elicitation

    National Research Council Canada - National Science Library

    Braun, Kimberly

    1997-01-01

    .... This thesis examines scenarios used in software requirements elicitation. Many different definitions, formats, and ideas exist on scenarios, but no thorough work has been done on what makes a good, quality scenario and scenario set...

  15. An integrated approach for requirement selection and scheduling in software release planning

    NARCIS (Netherlands)

    Li, C.; van den Akker, Marjan; Brinkkemper, Sjaak; Diepen, Guido

    2010-01-01

    It is essential for product software companies to decide which requirements should be included in the next release and to make an appropriate time plan of the development project. Compared to the extensive research done on requirement selection, very little research has been performed on time

  16. Requirements for flexible learner monitoring

    NARCIS (Netherlands)

    Glahn, Christian; Specht, Marcus; Koper, Rob

    2007-01-01

    Glahn, C., Specht, M., & Koper, R. (2007). Requirements for flexible learner monitoring. In T. Navarette, J. Blat & R. Koper (Eds.). Proceedings of the 3rd TENCompetence Open Workshop 'Current Research on IMS Learning Design and Lifelong Competence Development Infrastructures' (pp. 89-96). June,

  17. System requirements for one-time-use ENRAF control panel software

    International Nuclear Information System (INIS)

    HUBER, J.H.

    1999-01-01

    An Enraf Densitometer is installed on tank 241-AY-102. The Densitometer will frequently be tasked to obtain and log density profiles. The activity can be effected a number of ways. Enraf Incorporated provides a software package called ''Logger18'' to its customers for the purpose of in-shop testing of their gauges. Logger18 is capable of accepting an input file which can direct the gauge to obtain a density profile for a given tank level and bottom limit. Logger18 is a complex, DOS based program which will require trained technicians and/or tank farm entries to obtain the data. ALARA considerations have prompted the development of a more user-friendly, computer-based interface to the Enraf densitometers. This document records the plan by which this new Enraf data acquisition software will be developed, reviewed, verified, and released. This plan applies to the development and implementation of a one-time-use software program, which will be called ''Enraf Control Panel.'' The software will be primarily used for remote operation of Enraf Densitometers for the purpose of obtaining and logging tank product density profiles

  18. OncoTREAT: a software assistant for cancer therapy monitoring

    International Nuclear Information System (INIS)

    Bornemann, Lars; Dicken, Volker; Kuhnigk, Jan-Martin; Krass, Stefan; Peitgen, Heinz-Otto; Wormanns, Dag; Shin, Hoen-Oh; Bauknecht, Hans-Christian; Diehl, Volker; Fabel, Michael; Meier, Stefan; Kress, Oliver

    2007-01-01

    ObjectCancer is one of the leading causes of death worldwide and therapy options are often associated with severe stress for the patient and high costs. Therefore, precise evaluation of therapy success is essential. Material and Methods In the framework of the VICORA research project (Virtual Institute for Computer Assistance in Clinical Radiology), a software application was developed to support the radiologist in evaluating the response to tumor therapy. The application provides follow-up support for oncological therapy monitoring by volumetric quantification of lung, liver and brain metastases as well as enlarged lymph nodes and assists the user by temporal registration of lesion positions. Results With close cooperation between computer scientists and radiologists the application was tested and optimized to achieve a high degree of usability. Several clinical studies were carried out to evaluate the robustness and reproducibility of the volumetry methods. Conclusion Automatic volumetry and segmentation allows reliable detection of tumor growth and has the potential to increase reliability and significance of monitoring tumor growth in follow-up examinations. (orig.)

  19. National Ignition Facility subsystem design requirements supervisory control software SSDR 1.5.2

    International Nuclear Information System (INIS)

    Woodruff, J.; VanArsdall, P.; Bliss, E.

    1996-01-01

    This System Design Requirement document establishes the performance, design, development, and test requirements for the Supervisory Control Software, WBS 1.5.2, which is part of the NIF Integrated Computer Control System (ICCS). This document responds directly to the requirements detailed in ICCS (WBS 1-5)

  20. The Computer-based Health Evaluation Software (CHES: a software for electronic patient-reported outcome monitoring

    Directory of Open Access Journals (Sweden)

    Holzner Bernhard

    2012-11-01

    Full Text Available Abstract Background Patient-reported Outcomes (PROs capturing e.g., quality of life, fatigue, depression, medication side-effects or disease symptoms, have become important outcome parameters in medical research and daily clinical practice. Electronic PRO data capture (ePRO with software packages to administer questionnaires, storing data, and presenting results has facilitated PRO assessment in hospital settings. Compared to conventional paper-pencil versions of PRO instruments, ePRO is more economical with regard to staff resources and time, and allows immediate presentation of results to the medical staff. The objective of our project was to develop software (CHES – Computer-based Health Evaluation System for ePRO in hospital settings and at home with a special focus on the presentation of individual patient’s results. Methods Following the Extreme Programming development approach architecture was not fixed up-front, but was done in close, continuous collaboration with software end users (medical staff, researchers and patients to meet their specific demands. Developed features include sophisticated, longitudinal charts linking patients’ PRO data to clinical characteristics and to PRO scores from reference populations, a web-interface for questionnaire administration, and a tool for convenient creating and editing of questionnaires. Results By 2012 CHES has been implemented at various institutions in Austria, Germany, Switzerland, and the UK and about 5000 patients participated in ePRO (with around 15000 assessments in total. Data entry is done by the patients themselves via tablet PCs with a study nurse or an intern approaching patients and supervising questionnaire completion. Discussion During the last decade several software packages for ePRO have emerged for different purposes. Whereas commercial products are available primarily for ePRO in clinical trials, academic projects have focused on data collection and presentation in daily

  1. Development and implementation of a software for assessment of calculations in radiotherapy monitor units

    International Nuclear Information System (INIS)

    Mueller, Marcio Rogerio; Silva, Marco Antonio da; Rodrigues, Laura Natal

    2005-01-01

    Objective: To develop and implement a software based on the manual calculation algorithm for verification of the calculations in radiotherapy monitor units and to establish acceptance levels as a mechanism of quality assurance. Materials and methods: Data were obtained from Clinac 600C and 2100C Varian linear accelerators and the computerized treatment planning system used was CadPlan TM . Results: For the 6 MV beams the acceptance levels for deviations among the calculations of monitor units, divided by treatment area, were the following: breast (0.0%±1.7%), head and neck (1.5%±0.5%), hypophysis (-1.7%±0.5%), pelvis (2.1%±2.1%) and thorax (0.2%±1.3%). For the 15 MV beams the suggested level for pelvis in all the treatment techniques was (3.2%±1.3%). Conclusion: The present data are sufficient to justify the use of the software in the clinical practice as a tool for the quality assurance program. (author)

  2. VETRA - offline analysis and monitoring software platform for the LHCb Vertex Locator

    International Nuclear Information System (INIS)

    Szumlak, Tomasz

    2010-01-01

    The LHCb experiment is dedicated to studying CP violation and rare decay phenomena. In order to achieve these physics goals precise tracking and vertexing around the interaction point is crucial. This is provided by the VELO (VErtex LOcator) silicon detector. After digitization, FPGAs are employed to run several algorithms to suppress noise and reconstruct clusters. This is performed by an FPGA based processing board. An off-line software project, VETRA, has been developed which performs a bit perfect emulation of this complex processing in the FPGAs. This is a novel development as this hardware emulation is not standalone but rather is fully integrated into the LHCb software to allow the reconstruction of full data from the detector. This software platform facilitates the development and understanding of the behaviour of the processing algorithms, the optimization of the parameters of the algorithms that will be loaded into the FPGA and monitoring of the detector performance. This framework has also been adopted by the Silicon Tracker detector of LHCb. This processing framework was successfully used with the first 1500 tracks of data in the VELO obtained from the first LHC beam in September 2008. The software architecture and utilisation of the VETRA project will be discussed in detail.

  3. Monitored Retrievable Storage System Requirements Document

    International Nuclear Information System (INIS)

    1994-03-01

    This Monitored Retrievable Storage System Requirements Document (MRS-SRD) describes the functions to be performed and technical requirements for a Monitored Retrievable Storage (MRS) facility subelement and the On-Site Transfer and Storage (OSTS) subelement. The MRS facility subelement provides for temporary storage, at a Civilian Radioactive Waste Management System (CRWMS) operated site, of spent nuclear fuel (SNF) contained in an NRC-approved Multi-Purpose Canister (MPC) storage mode, or other NRC-approved storage modes. The OSTS subelement provides for transfer and storage, at Purchaser sites, of spent nuclear fuel (SNF) contained in MPCs. Both the MRS facility subelement and the OSTS subelement are in support of the CRWMS. The purpose of the MRS-SRD is to define the top-level requirements for the development of the MRS facility and the OSTS. These requirements include design, operation, and decommissioning requirements to the extent they impact on the physical development of the MRS facility and the OSTS. The document also presents an overall description of the MRS facility and the OSTS, their functions (derived by extending the functional analysis documented by the Physical System Requirements (PSR) Store Waste Document), their segments, and the requirements allocated to the segments. In addition, the top-level interface requirements of the MRS facility and the OSTS are included. As such, the MRS-SRD provides the technical baseline for the MRS Safety Analysis Report (SAR) design and the OSTS Safety Analysis Report design

  4. Automated Intelligent Monitoring and the Controlling Software System for Solar Panels

    Science.gov (United States)

    Nalamwar, H. S.; Ivanov, M. A.; Baidali, S. A.

    2017-01-01

    The inspection of the solar panels on a periodic basis is important to improve longevity and ensure performance of the solar system. To get the most solar potential of the photovoltaic (PV) system is possible through an intelligent monitoring & controlling system. The monitoring & controlling system has rapidly increased its popularity because of its user-friendly graphical interface for data acquisition, monitoring, controlling and measurements. In order to monitor the performance of the system especially for renewable energy source application such as solar photovoltaic (PV), data-acquisition systems had been used to collect all the data regarding the installed system. In this paper the development of a smart automated monitoring & controlling system for the solar panel is described, the core idea is based on IoT (the Internet of Things). The measurements of data are made using sensors, block management data acquisition modules, and a software system. Then, all the real-time data collection of the electrical output parameters of the PV plant such as voltage, current and generated electricity is displayed and stored in the block management. The proposed system is smart enough to make suggestions if the panel is not working properly, to display errors, to remind about maintenance of the system through email or SMS, and to rotate panels according to a sun position using the Ephemeral table that stored in the system. The advantages of the system are the performance of the solar panel system which can be monitored and analyzed.

  5. Investigation of the current requirements engineering practices among software developers at the Universiti Utara Malaysia Information Technology (UUMIT) centre

    Science.gov (United States)

    Hussain, Azham; Mkpojiogu, Emmanuel O. C.; Abdullah, Inam

    2016-08-01

    Requirements Engineering (RE) is a systemic and integrated process of eliciting, elaborating, negotiating, validating and managing of the requirements of a system in a software development project. UUM has been supported by various systems developed and maintained by the UUM Information Technology (UUMIT) Centre. The aim of this study was to assess the current requirements engineering practices at UUMIT. The main problem that prompted this research is the lack of studies that support software development activities at the UUMIT. The study is geared at helping UUMIT produce quality but time and cost saving software products by implementing cutting edge and state of the art requirements engineering practices. Also, the study contributes to UUM by identifying the activities needed for software development so that the management will be able to allocate budget to provide adequate and precise training for the software developers. Three variables were investigated: Requirement Description, Requirements Development (comprising: Requirements Elicitation, Requirements Analysis and Negotiation, Requirements Validation), and Requirement Management. The results from the study showed that the current practice of requirement engineering in UUMIT is encouraging, but still need further development and improvement because a few RE practices were seldom practiced.

  6. Requirements for internal contamination monitoring units

    International Nuclear Information System (INIS)

    Werner, E.; Beyer, D.; Doerfel, H.; Erlenbach, H.; Fischbach, G.; Henrichs, K.; Keller, K.D.; Koenig, K.; Riedel, W.; Scheler, R.; Schieferdecker, H.

    1994-01-01

    For the evaluation of internal contamination by officially authorized monitoring units according to section 63 (6) of the German radiation protection ordinance, a directive will be prepared in order to define uniform requirements for the laboratories in charge of incorporation monitoring by appointment of the relevant authorities. These requirements refer to equipment, to performance of measurements, to interpretation of measured values, to quality control as well as to documentation and to delivery of results. The duties of such laboratories comprise measurements of radioactivity, evaluation of intakes of radionuclides and of the resulting internal radiation dose, but also transmission of results to the employer, to the central dose registry, and under certain circumstances to the authorities. Among the requirements to be met by the laboratory for incorporation control are a sufficient measuring efficiency, the implementation of in-house checks, and the participation in intercomparison programs. For the accomplishment of such duties the laboratory needs appropriate apparatus, rooms, facilities, and staffing. (orig.) [de

  7. The Challenges of the "Software Support for Industrial Controls" Contract

    CERN Document Server

    Ninin, P

    2000-01-01

    ST division is currently specifying a 'Software Support for Industrial Controls' contract. The application of this contract and its success will require several changes in our habits for specifying, designing, and maintaining control systems. This paper summarizes some key concepts which should be respected in order to obtain maximum benefits from the future contract and to optimize the software activities in the division. The contract concerns the maintenance and development of the monitoring and control systems used for supervising CERN's technical infrastructure (electrical distribution, cooling water, air conditioning, safety, and access control). The systems concerned consist of computer and communication hardware and software, tailored to provide specific functionalities for the remote operation, command, and monitoring of equipment. All these systems use commercially available software and hardware such as SCADA, PLCs and associated drivers, controllers, fieldbuses, and networks. It is intended to cont...

  8. Requirement analysis of the safety-critical software implementation for the nuclear power plant

    International Nuclear Information System (INIS)

    Chang, Hoon Seon; Jung, Jae Cheon; Kim, Jae Hack; Nam, Sang Ku; Kim, Hang Bae

    2005-01-01

    The safety critical software shall be implemented under the strict regulation and standards along with hardware qualification. In general, the safety critical software has been implemented using functional block language (FBL) and structured language like C in the real project. Software design shall comply with such characteristics as; modularity, simplicity, minimizing the use of sub-routine, and excluding the interrupt logic. To meet these prerequisites, we used the computer-aided software engineering (CASE) tool to substantiate the requirements traceability matrix that were manually developed using Word processors or Spreadsheets. And the coding standard and manual have been developed to confirm the quality of software development process, such as; readability, consistency, and maintainability in compliance with NUREG/CR-6463. System level preliminary hazard analysis (PHA) is performed by analyzing preliminary safety analysis report (PSAR) and FMEA document. The modularity concept is effectively implemented for the overall module configurations and functions using RTP software development tool. The response time imposed on the basis of the deterministic structure of the safety-critical software was measured

  9. Requirements Specification for Open Source Software Selection

    OpenAIRE

    YANG, YING

    2008-01-01

    Open source software has been widely used. The software world is enjoying the advantages of collaboration and cooperation in software development and use with the advent of open source movement. However, little research is concerned about the practical guidelines of OSS selection. It is hard for an organization to make a decision whether they should use the OSS or not, and to select an appropriate one from a number of OSS candidates. This thesis studies how to select an open source software f...

  10. Monitoring of computing resource use of active software releases at ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00219183; The ATLAS collaboration

    2017-01-01

    The LHC is the world’s most powerful particle accelerator, colliding protons at centre of mass energy of 13 TeV. As the energy and frequency of collisions has grown in the search for new physics, so too has demand for computing resources needed for event reconstruction. We will report on the evolution of resource usage in terms of CPU and RAM in key ATLAS offline reconstruction workflows at the TierO at CERN and on the WLCG. Monitoring of workflows is achieved using the ATLAS PerfMon package, which is the standard ATLAS performance monitoring system running inside Athena jobs. Systematic daily monitoring has recently been expanded to include all workflows beginning at Monte Carlo generation through to end-user physics analysis, beyond that of event reconstruction. Moreover, the move to a multiprocessor mode in production jobs has facilitated the use of tools, such as “MemoryMonitor”, to measure the memory shared across processors in jobs. Resource consumption is broken down into software domains and dis...

  11. 40 CFR 146.13 - Operating, monitoring and reporting requirements.

    Science.gov (United States)

    2010-07-01

    ...) Monitoring requirements. Monitoring requirements shall, at a minimum, include: (1) The analysis of the... minimum, include: (1) Quarterly reports to the Director on: (i) The physical, chemical and other relevant... in the first aquifer overlying the confining zone. When such a well is installed, the owner or...

  12. Design of a Nanosatellite Ground Monitoring and Control Software – a Case Study

    Directory of Open Access Journals (Sweden)

    Freddy Alexander Díaz González

    2016-04-01

    Full Text Available The growing countries that have carried out the development of CubeSat missions for academic purposes do not offer aerospace engineering programs at their universities. This causes difficulties for traditional engineers upon the formal use of different standards and frameworks for aerospace development, such as the European Cooperation for Space Standardization and Space Mission Analysis and Design . One way in which traditional software engineers can easily understand the structure of an aerospace framework, in order to apply it on the development of CubeSat mission software parts, is comparing its most important elements in relation to the elements suggested by a more familiar method. In this paper, we present a hybrid framework between the ECSS-E-ST-40C standard and the Rational Unified Process, which can be used by traditional software engineers as a guide model for the development of software elements in academic nanosatellite missions. The model integrates the processes and documentation suggested by the ECSS-E-ST-40C with the disciplines, workflows and artifacts suggested in Rational Unified Process. This simplifies the structure of ECSS-E-ST-40C and allows traditional software engineers to easily understand its work elements. The paper describes as study case the implementation of the hybrid model in the analysis and design of ground monitoring and control software for the Libertad-2 satellite mission, which is currently being developed by the Universidad Sergio Arboleda in Colombia.

  13. Using iKidTools™ Software Support Systems to Develop and Implement Self-Monitoring Interventions

    Science.gov (United States)

    Patti, Angela L.; Miller, Kevin J.

    2011-01-01

    Educational teams often are faced with the task of developing and implementing Behavioral Intervention Plans (BIPs) for students who present challenging and/or disruptive behaviors. This article describes the steps used to develop and implement a self-monitoring BIP that incorporated an innovative software system, iKidTools™. An authentic case…

  14. Improvement of Computer Software Quality through Software Automated Tools.

    Science.gov (United States)

    1986-08-30

    information that are returned from the tools to the human user, and the forms in which these outputs are presented. Page 2 of 4 STAGE OF DEVELOPMENT: What... AUTOMIATED SOFTWARE TOOL MONITORING SYSTEM APPENDIX 2 2-1 INTRODUCTION This document and Automated Software Tool Monitoring Program (Appendix 1) are...t Output Output features provide links from the tool to both the human user and the target machine (where applicable). They describe the types

  15. A software engineering process for safety-critical software application

    International Nuclear Information System (INIS)

    Kang, Byung Heon; Kim, Hang Bae; Chang, Hoon Seon; Jeon, Jong Sun

    1995-01-01

    Application of computer software to safety-critical systems in on the increase. To be successful, the software must be designed and constructed to meet the functional and performance requirements of the system. For safety reason, the software must be demonstrated not only to meet these requirements, but also to operate safely as a component within the system. For longer-term cost consideration, the software must be designed and structured to ease future maintenance and modifications. This paper presents a software engineering process for the production of safety-critical software for a nuclear power plant. The presentation is expository in nature of a viable high quality safety-critical software development. It is based on the ideas of a rational design process and on the experience of the adaptation of such process in the production of the safety-critical software for the shutdown system number two of Wolsung 2, 3 and 4 nuclear power generation plants. This process is significantly different from a conventional process in terms of rigorous software development phases and software design techniques, The process covers documentation, design, verification and testing using mathematically precise notations and highly reviewable tabular format to specify software requirements and software requirements and software requirements and code against software design using static analysis. The software engineering process described in this paper applies the principle of information-hiding decomposition in software design using a modular design technique so that when a change is required or an error is detected, the affected scope can be readily and confidently located. it also facilitates a sense of high degree of confidence in the 'correctness' of the software production, and provides a relatively simple and straightforward code implementation effort. 1 figs., 10 refs. (Author)

  16. Commercial Literacy Software.

    Science.gov (United States)

    Balajthy, Ernest

    1997-01-01

    Presents the first year's results of a continuing project to monitor the availability of software of relevance for literacy education purposes. Concludes there is an enormous amount of software available for use by teachers of reading and literacy--whereas drill-and-practice software is the largest category of software available, large numbers of…

  17. Advanced Modular Software Performance Monitoring

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    The LHCb software is based on the Gaudi framework, on top of which are built several large and complex software applications. The LHCb experiment is now in the active phase of collecting and analyzing data and significant performance problems arise in the Gaudi based software beginning from High Level Trigger (HLT) programs and ending with data analysis frameworks (DaVinci). It’s not easy to find hot spots in the code - only special tools can help to understand where CPU or memory usage is not reasonable. There exist many performance analyzing tools, but the main problem is that they show reports in terms of class and function names and such information usually is not very useful - the majority of algorithm developers use the Gaudi framework abstractions and usually do not know about functions which lie at the lower level. We will show a new approach which adds to performance reports a higher abstraction level based on knowledge of framework architecture and run-time object properties. A set of profiling to...

  18. Advanced modular software performance monitoring

    CERN Document Server

    Mazurov, A

    2012-01-01

    The LHCb software is based on the Gaudi framework, on top of which are built several large and complex software applications. As the LHCb experiment is now in the active phase of collecting and analyzing data, performance problems arise in various parts of the software, from the High Level Trigger (HLT) programs to data analysis frameworks. It is not easy to find hotspots in the code - only specialized tools can help to understand where CPU or memory usage are not reasonable. There exist many performance analyzing tools, but the main problem is that they show reports in terms of class and function names and such information usually is not very useful - the majority of algorithm developers use the Gaudi framework abstractions and usually do not know about functions which lie at the lower level. We will show a new approach which adds to performance reports a higher abstraction level based on knowledge of framework architecture and run-time object properties. A set of profiling tools (based on Intel VTune Amplif...

  19. Formalization of software requirements for information systems using fuzzy logic

    Science.gov (United States)

    Yegorov, Y. S.; Milov, V. R.; Kvasov, A. S.; Sorokoumova, S. N.; Suvorova, O. V.

    2018-05-01

    The paper considers an approach to the design of information systems based on flexible software development methodologies. The possibility of improving the management of the life cycle of information systems by assessing the functional relationship between requirements and business objectives is described. An approach is proposed to establish the relationship between the degree of achievement of business objectives and the fulfillment of requirements for the projected information system. It describes solutions that allow one to formalize the process of formation of functional and non-functional requirements with the help of fuzzy logic apparatus. The form of the objective function is formed on the basis of expert knowledge and is specified via learning from very small data set.

  20. A discussion of higher order software concepts as they apply to functional requirements and specifications. [space shuttles and guidance

    Science.gov (United States)

    Hamilton, M.

    1973-01-01

    The entry guidance software functional requirements (requirements design phase), its architectural requirements (specifications design phase), and the entry guidance software verified code are discussed. It was found that the proper integration of designs at both the requirements and specifications levels are of high priority consideration.

  1. Reliability of adaptive multivariate software sensors for sewer water quality monitoring

    DEFF Research Database (Denmark)

    Vezzaro, Luca; Mikkelsen, Peter Steen

    2015-01-01

    This study investigates the use of a multivariate approach, based on Principal Component Analysis PCA), as software sensor for fault detection and reconstruction of missing measurements in on-line monitoring of sewer water quality. The analysis was carried out on a 16-months dataset of five...... commonly available on-line measurements (flow, turbidity, ammonia, conductivity and temperature). The results confirmed the great performance of PCA (up to 10 weeks after parameter estimation) when estimating a measurement from the combination of the remaining four variables, a useful feature in data...

  2. Software Validation in ATLAS

    International Nuclear Information System (INIS)

    Hodgkinson, Mark; Seuster, Rolf; Simmons, Brinick; Sherwood, Peter; Rousseau, David

    2012-01-01

    The ATLAS collaboration operates an extensive set of protocols to validate the quality of the offline software in a timely manner. This is essential in order to process the large amounts of data being collected by the ATLAS detector in 2011 without complications on the offline software side. We will discuss a number of different strategies used to validate the ATLAS offline software; running the ATLAS framework software, Athena, in a variety of configurations daily on each nightly build via the ATLAS Nightly System (ATN) and Run Time Tester (RTT) systems; the monitoring of these tests and checking the compilation of the software via distributed teams of rotating shifters; monitoring of and follow up on bug reports by the shifter teams and periodic software cleaning weeks to improve the quality of the offline software further.

  3. Software quality assurance plans for safety-critical software

    International Nuclear Information System (INIS)

    Liddle, P.

    2006-01-01

    Application software is defined as safety-critical if a fault in the software could prevent the system components from performing their nuclear-safety functions. Therefore, for nuclear-safety systems, the AREVA TELEPERM R XS (TXS) system is classified 1E, as defined in the Inst. of Electrical and Electronics Engineers (IEEE) Std 603-1998. The application software is classified as Software Integrity Level (SIL)-4, as defined in IEEE Std 7-4.3.2-2003. The AREVA NP Inc. Software Program Manual (SPM) describes the measures taken to ensure that the TELEPERM XS application software attains a level of quality commensurate with its importance to safety. The manual also describes how TELEPERM XS correctly performs the required safety functions and conforms to established technical and documentation requirements, conventions, rules, and standards. The program manual covers the requirements definition, detailed design, integration, and test phases for the TELEPERM XS application software, and supporting software created by AREVA NP Inc. The SPM is required for all safety-related TELEPERM XS system applications. The program comprises several basic plans and practices: 1. A Software Quality-Assurance Plan (SQAP) that describes the processes necessary to ensure that the software attains a level of quality commensurate with its importance to safety function. 2. A Software Safety Plan (SSP) that identifies the process to reasonably ensure that safety-critical software performs as intended during all abnormal conditions and events, and does not introduce any new hazards that could jeopardize the health and safety of the public. 3. A Software Verification and Validation (V and V) Plan that describes the method of ensuring the software is in accordance with the requirements. 4. A Software Configuration Management Plan (SCMP) that describes the method of maintaining the software in an identifiable state at all times. 5. A Software Operations and Maintenance Plan (SO and MP) that

  4. Hardware and software concept of the remote monitoring system for nuclear reactors in Bavaria

    International Nuclear Information System (INIS)

    Gietl, G.

    1981-01-01

    The remote monitoring system for nuclear reactors (KFUe) is a fully automatic system for the measuring and registration of radioactive releases and of operation parameters of the nuclear power plants in Bavaria. It has also to measure the meteorological parameters on the site of a power plant for the purpose of dispersion calculations. The system consists of the network centre and the subsystems with their satellite stations at the nuclear power plants. Process computers in the network centre and in the subsystems perform the automatic operation and they handle the data transmission via data lines. The hardware of the entire system is so conceived that the mesured data can be processed, stored and displayed in an optimal way for the user. The software of the KFUe centre and the subsystems is structured strictly modular. Redunancies are implied in the hardware and software. The automatic remote air quality monitoring system of Bavaria (LUeB) can be operated as a stand-by system for the KFUe and vice versa in case of an emergency. (orig./HP) [de

  5. Implementation of electronic medical records requires more than new software: Lessons on integrating and managing health technologies from Mbarara, Uganda.

    Science.gov (United States)

    Madore, Amy; Rosenberg, Julie; Muyindike, Winnie R; Bangsberg, David R; Bwana, Mwebesa B; Martin, Jeffrey N; Kanyesigye, Michael; Weintraub, Rebecca

    2015-12-01

    Implementation lessons: • Technology alone does not necessarily lead to improvement in health service delivery, in contrast to the common assumption that advanced technology goes hand in hand with progress. • Implementation of electronic medical record (EMR) systems is a complex, resource-intensive process that, in addition to software, hardware, and human resource investments, requires careful planning, change management skills, adaptability, and continuous engagement of stakeholders. • Research requirements and goals must be balanced with service delivery needs when determining how much information is essential to collect and who should be interfacing with the EMR system. • EMR systems require ongoing monitoring and regular updates to ensure they are responsive to evolving clinical use cases and research questions. • High-quality data and analyses are essential for EMRs to deliver value to providers, researchers, and patients. Copyright © 2015 Elsevier Inc. All rights reserved.

  6. Fully Employing Software Inspections Data

    Science.gov (United States)

    Shull, Forrest; Feldmann, Raimund L.; Seaman, Carolyn; Regardie, Myrna; Godfrey, Sally

    2009-01-01

    Software inspections provide a proven approach to quality assurance for software products of all kinds, including requirements, design, code, test plans, among others. Common to all inspections is the aim of finding and fixing defects as early as possible, and thereby providing cost savings by minimizing the amount of rework necessary later in the lifecycle. Measurement data, such as the number and type of found defects and the effort spent by the inspection team, provide not only direct feedback about the software product to the project team but are also valuable for process improvement activities. In this paper, we discuss NASA's use of software inspections and the rich set of data that has resulted. In particular, we present results from analysis of inspection data that illustrate the benefits of fully utilizing that data for process improvement at several levels. Examining such data across multiple inspections or projects allows team members to monitor and trigger cross project improvements. Such improvements may focus on the software development processes of the whole organization as well as improvements to the applied inspection process itself.

  7. PDSS/IMC requirements and functional specifications

    Science.gov (United States)

    1983-01-01

    The system (software and hardware) requirements for the Payload Development Support System (PDSS)/Image Motion Compensator (IMC) are provided. The PDSS/IMC system provides the capability for performing Image Motion Compensator Electronics (IMCE) flight software test, checkout, and verification and provides the capability for monitoring the IMC flight computer system during qualification testing for fault detection and fault isolation.

  8. Remote monitoring of videourodynamics using smart phone and free instant messaging software.

    Science.gov (United States)

    Hsieh, Po-Fan; Chang, Chao-Hsiang; Lien, Chi-Shun; Wu, Hsi-Chin; Hsiao, Po-Jen; Chou, Eric Chieh-Lung

    2013-11-01

    To evaluate the feasibility of using smart phones plus free instant messaging software for remote monitoring of videourodynamics. From November 2011 to October 2012, 85 females with voiding disorders were enrolled for videourodynamic tests. The patients were assigned to videourodynamics remotely monitored by the attending physician by using iPhone/iPad and Skype (group 1) and videourodynamics with the attending physician present (group 2). The procedural time and videourodynamic qualities, assessed by the frequency of adherence to the modified Sullivan criteria, in each group were recorded and compared. There were 44 and 41 patients in group 1 and group 2, respectively. The mean procedural time was comparable between group 1 and group 2 (56.3 vs. 54.4 min, P = 0.25). The frequencies of adherence to the modified Sullivan criteria were similar in each group. The qualities of videourodynamics under the attending physician's remote or direct monitoring were both appropriate. Based on the convenience of Internet, the popularity of smart phones and the intention to make the urologists use their time more efficiently, our study provides remote monitoring as an alternative way for performing videourodynamics. © 2013 Wiley Periodicals, Inc.

  9. Risk monitor riskangel for risk-informed applications in nuclear power plants

    International Nuclear Information System (INIS)

    Wang, Fang; Wang, Jiaqun; Wang, Jin; Li, Yazhou; Hu, Liqin; Wu, Yican

    2016-01-01

    Highlights: • A general risk monitor riskangel with high-speed cutsets generator engine. • Benchmarks of actual nuclear power plant (NPP) instantaneous risk models. • Applications in daily operation, maintenance plan and component out of service. - Abstract: This paper studied the requirements of risk monitor software and its applications as a plant specific risk monitor, which supports risk-informed configuration risk management for the two CANDU 6 units at the Third Qinshan nuclear power plant (TQNPP) in China. It also describes the regulatory prospective on risk-informed Probabilistic Safety Assessment (PSA) applications and the use of risk monitor at operating nuclear power plants, high level technical and functional requirements for the development of CANDU specific risk monitor software, and future development trends.

  10. Basic requirements for personnel monitoring. 1980 ed

    International Nuclear Information System (INIS)

    1980-01-01

    This Code of Practice sets forth the objectives of an adequate system of personnel monitoring for radiation workers. It covers individual dosimetry, including internal radiation monitoring, and area monitoring to the extent required for the assessment of individual radiation doses. The responsibilities of authorities for organizing monitoring of radiation workers are discussed, together with brief descriptions of monitoring methods and the rules governing their application. The general principles to be considered in selecting instrumentation and the appropriate monitoring techniques are described, as well as calibration techniques, methods of data handling and record keeping. Current concepts and recommendations of the International Commission on Radiological Protection, as presented in ICRP Publication No.26, have been incorporated. New developments in techniques and instruments have been reflected, and several sections such as calibration and record keeping have been elaborated. The bibliography has been updated and new annexes added.

  11. An Industrial Case Study in Reconstructing Requirements Views

    NARCIS (Netherlands)

    Lormans, M.; Van Deursen, A.; Gross, H.G.

    2008-01-01

    Preprint of article published in: Empirical Software Engineering (Springer), 13 (6), 2008; doi:10.1007/s10664-008-9078-4 Requirements views, such as coverage and status views, are an important asset for monitoring and managing software development projects. We have developed a method that automates

  12. 40 CFR 425.06 - Monitoring requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 29 2010-07-01 2010-07-01 false Monitoring requirements. 425.06 Section 425.06 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) EFFLUENT GUIDELINES AND STANDARDS LEATHER TANNING AND FINISHING POINT SOURCE CATEGORY General Provisions § 425.06...

  13. 50 CFR 660.15 - Equipment requirements.

    Science.gov (United States)

    2010-10-01

    ... software requirements. (i) A personal computer with Pentium 75-MHz or higher. Random Access Memory (RAM... Graphics Adapter (VGA) or higher resolution monitor (super VGA is recommended). (ii) Microsoft Windows 2000...

  14. Requirements to a Norwegian National Automatic Gamma Monitoring System

    International Nuclear Information System (INIS)

    Lauritzen, B.; Hedemann Jensen, P.; Nielsen, F.

    2005-04-01

    An assessment of the overall requirements to a Norwegian gamma-monitoring network is undertaken with special emphasis on the geographical distribution of automatic gamma monitoring stations, type of detectors in such stations and the sensitivity of the system in terms of ambient dose equivalent rate increments above the natural background levels. The study is based upon simplified deterministic calculations of the radiological consequences of generic nuclear accident scenarios. The density of gamma monitoring stations has been estimated from an analysis of the dispersion of radioactive materials over large distances using historical weather data; the minimum density is estimated from the requirement that a radioactive plume may not slip unnoticed in between stations of the monitoring network. The sensitivity of the gamma monitoring system is obtained from the condition that events that may require protective intervention measures should be detected by the system. Action levels for possible introduction of sheltering and precautionary foodstuff restrictions are derived in terms of ambient dose equivalent rate. For emergency situations where particulates contribute with only a small fraction of the total ambient dose equivalent rate from the plume, it is concluded that measurements of dose rate are sufficient to determine the need for sheltering; simple dose rate measurements however, are inadequate to determine the need for foodstuff restrictions and spectral measurements are required. (au)

  15. Detailed requirements document for common software of shuttle program information management system

    Science.gov (United States)

    Everette, J. M.; Bradfield, L. D.; Horton, C. L.

    1975-01-01

    Common software was investigated as a method for minimizing development and maintenance cost of the shuttle program information management system (SPIMS) applications while reducing the time-frame of their development. Those requirements satisfying these criteria are presented along with the stand-alone modules which may be used directly by applications. The SPIMS applications operating on the CYBER 74 computer, are specialized information management systems which use System 2000 as a data base manager. Common software provides the features to support user interactions on a CRT terminal using form input and command response capabilities. These features are available as subroutines to the applications.

  16. S-Cube: Enabling the Next Generation of Software Services

    Science.gov (United States)

    Metzger, Andreas; Pohl, Klaus

    The Service Oriented Architecture (SOA) paradigm is increasingly adopted by industry for building distributed software systems. However, when designing, developing and operating innovative software services and servicebased systems, several challenges exist. Those challenges include how to manage the complexity of those systems, how to establish, monitor and enforce Quality of Service (QoS) and Service Level Agreements (SLAs), as well as how to build those systems such that they can proactively adapt to dynamically changing requirements and context conditions. Developing foundational solutions for those challenges requires joint efforts of different research communities such as Business Process Management, Grid Computing, Service Oriented Computing and Software Engineering. This paper provides an overview of S-Cube, the European Network of Excellence on Software Services and Systems. S-Cube brings together researchers from leading research institutions across Europe, who join their competences to develop foundations, theories as well as methods and tools for future service-based systems.

  17. Effects of Using Requirements Catalogs on Effectiveness and Productivity of Requirements Specification in a Software Project Management Course

    Science.gov (United States)

    Fernández-Alemán, José Luis; Carrillo-de-Gea, Juan Manuel; Meca, Joaquín Vidal; Ros, Joaquín Nicolás; Toval, Ambrosio; Idri, Ali

    2016-01-01

    This paper presents the results of two educational experiments carried out to determine whether the process of specifying requirements (catalog-based reuse as opposed to conventional specification) has an impact on effectiveness and productivity in co-located and distributed software development environments. The participants in the experiments…

  18. 40 CFR 413.03 - Monitoring requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 28 2010-07-01 2010-07-01 true Monitoring requirements. 413.03 Section 413.03 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) EFFLUENT GUIDELINES AND... for total toxic organics (TTO), I certify that, to the best of my knowledge and belief, no dumping of...

  19. 40 CFR 437.4 - Monitoring requirements.

    Science.gov (United States)

    2010-07-01

    ... wastewater resulting from the treatment of metal-bearing waste, oily waste, or organic-bearing waste must... STANDARDS THE CENTRALIZED WASTE TREATMENT POINT SOURCE CATEGORY § 437.4 Monitoring requirements. (a) Permit... compliance for each subpart after treatment and before mixing of the waste with wastes of any other subpart...

  20. Outsourcing the development of specific application software using the ESA software engineering standards the SPS software Interlock System

    CERN Document Server

    Denis, B

    1995-01-01

    CERN is considering outsourcing as a solution to the reduction of staff. To need to re-engineer the SPS Software Interlock System provided an opportunity to explore the applicability of outsourcing to our specific controls environment and the ESA PSS-05 standards were selected for the requirements specification, the development, the control and monitoring and the project management. The software produced by the contractor is now fully operational. After outlining the scope and the complexity of the project, a discussion on the ESA PSS-05 will be presented: the choice, the way these standards improve the outsourcing process, the quality induced but also the need to adapt them and their limitation in the definition of the customer-supplier relationship. The success factors and the difficulties of development under contract will also be discussed. The maintenance aspect and the impact on in-house developments will finally be addressed.

  1. SCaN Testbed Software Development and Lessons Learned

    Science.gov (United States)

    Kacpura, Thomas J.; Varga, Denise M.

    2012-01-01

    National Aeronautics and Space Administration (NASA) has developed an on-orbit, adaptable, Software Defined Radio (SDR)Space Telecommunications Radio System (STRS)-based testbed facility to conduct a suite of experiments to advance technologies, reduce risk, and enable future mission capabilities on the International Space Station (ISS). The SCAN Testbed Project will provide NASA, industry, other Government agencies, and academic partners the opportunity to develop and field communications, navigation, and networking technologies in the laboratory and space environment based on reconfigurable, SDR platforms and the STRS Architecture.The SDRs are a new technology for NASA, and the support infrastructure they require is different from legacy, fixed function radios. SDRs offer the ability to reconfigure on-orbit communications by changing software for new waveforms and operating systems to enable new capabilities or fix any anomalies, which was not a previous option. They are not stand alone devices, but required a new approach to effectively control them and flow data. This requires extensive software to be developed to utilize the full potential of these reconfigurable platforms. The paper focuses on development, integration and testing as related to the avionics processor system, and the software required to command, control, monitor, and interact with the SDRs, as well as the other communication payload elements. An extensive effort was required to develop the flight software and meet the NASA requirements for software quality and safety. The flight avionics must be radiation tolerant, and these processors have limited capability in comparison to terrestrial counterparts. A big challenge was that there are three SDRs onboard, and interfacing with multiple SDRs simultaneously complicatesd the effort. The effort also includes ground software, which is a key element for both the command of the payload, and displaying data created by the payload. The verification of

  2. Remote sensing time series analysis for crop monitoring with the SPIRITS software: new functionalities and use examples

    Directory of Open Access Journals (Sweden)

    Felix eRembold

    2015-07-01

    Full Text Available Monitoring crop and natural vegetation conditions is highly relevant, particularly in the food insecure areas of the world. Data from remote sensing image time series at high temporal and medium to low spatial resolution can assist this monitoring as they provide key information about vegetation status in near real-time over large areas. The Software for the Processing and Interpretation of Remotely sensed Image Time Series (SPIRITS is a stand-alone flexible analysis environment created to facilitate the processing and analysis of large image time series and ultimately for providing clear information about vegetation status in various graphical formats to crop production analysts and decision makers. In this paper we present the latest functional developments of SPIRITS and we illustrate recent applications. The main new developments include: HDF5 importer, Image re-projection, additional options for temporal Smoothing and Periodicity conversion, computation of a rainfall-based probability index (Standardized Precipitation Index for drought detection and extension of the Graph composer functionalities.In particular,. The examples of operational analyses are taken from several recent agriculture and food security monitoring reports and bulletins. We conclude with considerations on future SPIRITS developments also in view of the data processing requirements imposed by the coming generation of remote sensing products at high spatial and temporal resolution, such as those provided by the Sentinel sensors of the European Copernicus programme.

  3. Alignment of process compliance and monitoring requirements in dynamic business collaborations

    Science.gov (United States)

    Comuzzi, Marco

    2017-07-01

    Dynamic business collaborations are intrinsically characterised by change because processes can be distributed or outsourced and partners may be substituted by new ones with enhanced or different capabilities. In this context, compliance requirements management becomes particularly challenging. Partners in a collaboration may join and leave dynamically and tasks over which compliance requirements are specified may be consequently distributed or delegated to new partners. This article considers the issue of aligning compliance requirements in a dynamic business collaboration with the monitoring requirements induced on the collaborating partners when change occurs. We first provide a conceptual model of business collaborations and their compliance requirements, introducing the concept of monitoring capabilities induced by compliance requirements. Then, we present a set of mechanisms to ensure consistency between monitoring and compliance requirements in the presence of change, e.g. when tasks are delegated or backsourced in-house. We also discuss a set of metrics to evaluate the status of a collaboration in respect of compliance monitorability. Finally, we discuss a prototype implementation of our framework.

  4. NASA software documentation standard software engineering program

    Science.gov (United States)

    1991-01-01

    The NASA Software Documentation Standard (hereinafter referred to as Standard) can be applied to the documentation of all NASA software. This Standard is limited to documentation format and content requirements. It does not mandate specific management, engineering, or assurance standards or techniques. This Standard defines the format and content of documentation for software acquisition, development, and sustaining engineering. Format requirements address where information shall be recorded and content requirements address what information shall be recorded. This Standard provides a framework to allow consistency of documentation across NASA and visibility into the completeness of project documentation. This basic framework consists of four major sections (or volumes). The Management Plan contains all planning and business aspects of a software project, including engineering and assurance planning. The Product Specification contains all technical engineering information, including software requirements and design. The Assurance and Test Procedures contains all technical assurance information, including Test, Quality Assurance (QA), and Verification and Validation (V&V). The Management, Engineering, and Assurance Reports is the library and/or listing of all project reports.

  5. Basic requirements of dosemeter systems for individual monitoring of external radiation

    International Nuclear Information System (INIS)

    Boehm, J.; Ambrosi, P.

    1985-01-01

    A plea is made for detailed detector independent requirements for dosemeter systems for individual monitoring of external radiation. These requirements should have their origin in the fundamental aspects and concepts of radiation protection for workers, and should be something like a translation of the general principles of individual monitoring into a language easily understandable by producers and users. This work comprises a summary of the general objectives of individual monitoring and discussion of some relevant requirements for dosemeter systems. (orig.) [de

  6. Supplemental Assessment of the Y-12 Groundwater Protection Program Using Monitoring and Remediation Optimization System Software

    Energy Technology Data Exchange (ETDEWEB)

    Elvado Environmental LLC; GSI Environmental LLC

    2009-01-01

    A supplemental quantitative assessment of the Groundwater Protection Program (GWPP) at the Y-12 National Security Complex (Y-12) in Oak Ridge, TN was performed using the Monitoring and Remediation Optimization System (MAROS) software. This application was previously used as part of a similar quantitative assessment of the GWPP completed in December 2005, hereafter referenced as the 'baseline' MAROS assessment (BWXT Y-12 L.L.C. [BWXT] 2005). The MAROS software contains modules that apply statistical analysis techniques to an existing GWPP analytical database in conjunction with hydrogeologic factors, regulatory framework, and the location of potential receptors, to recommend an improved groundwater monitoring network and optimum sampling frequency for individual monitoring locations. The goal of this supplemental MAROS assessment of the Y-12 GWPP is to review and update monitoring network optimization recommendations resulting from the 2005 baseline report using data collected through December 2007. The supplemental MAROS assessment is based on the findings of the baseline MAROS assessment and includes only the groundwater sampling locations (wells and natural springs) currently granted 'Active' status in accordance with the Y-12 GWPP Monitoring Optimization Plan (MOP). The results of the baseline MAROS assessment provided technical rationale regarding the 'Active' status designations defined in the MOP (BWXT 2006). One objective of the current report is to provide a quantitative review of data collected from Active but infrequently sampled wells to confirm concentrations at these locations. This supplemental MAROS assessment does not include the extensive qualitative evaluations similar to those presented in the baseline report.

  7. Unobtrusive Software and System Health Management with R2U2 on a Parallel MIMD Coprocessor

    Science.gov (United States)

    Schumann, Johann; Moosbrugger, Patrick

    2017-01-01

    Dynamic monitoring of software and system health of a complex cyber-physical system requires observers that continuously monitor variables of the embedded software in order to detect anomalies and reason about root causes. There exists a variety of techniques for code instrumentation, but instrumentation might change runtime behavior and could require costly software re-certification. In this paper, we present R2U2E, a novel realization of our real-time, Realizable, Responsive, and Unobtrusive Unit (R2U2). The R2U2E observers are executed in parallel on a dedicated 16-core EPIPHANY co-processor, thereby avoiding additional computational overhead to the system under observation. A DMA-based shared memory access architecture allows R2U2E to operate without any code instrumentation or program interference.

  8. Graph Based Verification of Software Evolution Requirements

    NARCIS (Netherlands)

    Ciraci, S.

    2009-01-01

    Due to market demands and changes in the environment, software systems have to evolve. However, the size and complexity of the current software systems make it time consuming to incorporate changes. During our collaboration with the industry, we observed that the developers spend much time on the

  9. The application of formal software engineering methods to the unattended and remote monitoring software suite at Los Alamos National Laboratory

    Energy Technology Data Exchange (ETDEWEB)

    Determan, John Clifford [Los Alamos National Laboratory; Longo, Joseph F [Los Alamos National Laboratory; Michel, Kelly D [Los Alamos National Laboratory

    2009-01-01

    The Unattended and Remote Monitoring (UNARM) system is a collection of specialized hardware and software used by the International Atomic Energy Agency (IAEA) to institute nuclear safeguards at many nuclear facilities around the world. The hardware consists of detectors, instruments, and networked computers for acquiring various forms of data, including but not limited to radiation data, global position coordinates, camera images, isotopic data, and operator declarations. The software provides two primary functions: the secure and reliable collection of this data from the instruments and the ability to perform an integrated review and analysis of the disparate data sources. Several years ago the team responsible for maintaining the software portion of the UNARM system began the process of formalizing its operations. These formal operations include a configuration management system, a change control board, an issue tracking system, and extensive formal testing, for both functionality and reliability. Functionality is tested with formal test cases chosen to fully represent the data types and methods of analysis that will be commonly encountered. Reliability is tested with iterative, concurrent testing where up to five analyses are executed simultaneously for thousands of cycles. Iterative concurrent testing helps ensure that there are no resource conflicts or leaks when multiple system components are in use simultaneously. The goal of this work is to provide a high quality, reliable product, commensurate with the criticality of the application. Testing results will be presented that demonstrate that this goal has been achieved and the impact of the introduction of a formal software engineering framework to the UNARM product will be presented.

  10. MODULAR CAUSTIC SIDE SOLVENT EXTRACTION UNIT (MCU) GAMMA MONITORS SYSTEM FINAL REPORT

    Energy Technology Data Exchange (ETDEWEB)

    Casella, V

    2005-12-15

    The Department of Energy (DOE) selected Caustic-Side Solvent Extraction (CSSX) as the preferred technology for the removal of radioactive cesium from High-Level Waste (HLW) at the Savannah River Site (SRS). Before the full-scale Salt Waste Processing Facility (SWPF) becomes operational, the Closure Business Unit (CBU) plans to process a portion of dissolved saltcake waste through a Modular CSSX Unit (MCU). This work was derived from Technical Task Request SP-TTR-2004-00013, ''Gamma Monitor for MCU''. The deliverables for this task are the hardware and software for the gamma monitors and a report summarizing the testing and acceptance of this equipment for use in the MCU. Gamma-ray monitors are required to: (1) Measure the Cs-137 concentration in the decontaminated salt solution before entering the DSS (Decontaminated Salt Solution) Hold Tank, (2) Measure the Cs-137 concentration in the strip effluent before entering the Strip Effluent Hold Tank, (3) Verify proper operation of the solvent extraction system by verifying material balance within the process (The DSS Hold Tank Cs-137 concentration will be very low and the Cs-137 concentration in the Strip Effluent Hold Tank will be fifteen times higher than the Cs-137 concentration in the Feed Tank.) Sodium iodide monitors are used to measure the Cs-137 concentration in the piping before the DSS Hold tank, while GM monitors are used for Cs-137 measurements before the Strip Effluent Hold Tank. Tungsten shields were designed using Monte Carlo calculations and fabricated to reduce the process background radiation at the detector positions. These monitors were calibrated with NIST traceable standards that were specially made to be the same as the piping being monitored. Since this gamma ray monitoring system is unique, specially designed software was written and acceptance tested by Savannah River National Laboratory personnel. The software is a LabView-based application that serves as a unified

  11. MODULAR CAUSTIC SIDE SOLVENT EXTRACTION UNIT (MCU) GAMMA MONITORS SYSTEM FINAL REPORT

    International Nuclear Information System (INIS)

    Casella, V

    2005-01-01

    The Department of Energy (DOE) selected Caustic-Side Solvent Extraction (CSSX) as the preferred technology for the removal of radioactive cesium from High-Level Waste (HLW) at the Savannah River Site (SRS). Before the full-scale Salt Waste Processing Facility (SWPF) becomes operational, the Closure Business Unit (CBU) plans to process a portion of dissolved saltcake waste through a Modular CSSX Unit (MCU). This work was derived from Technical Task Request SP-TTR-2004-00013, ''Gamma Monitor for MCU''. The deliverables for this task are the hardware and software for the gamma monitors and a report summarizing the testing and acceptance of this equipment for use in the MCU. Gamma-ray monitors are required to: (1) Measure the Cs-137 concentration in the decontaminated salt solution before entering the DSS (Decontaminated Salt Solution) Hold Tank, (2) Measure the Cs-137 concentration in the strip effluent before entering the Strip Effluent Hold Tank, (3) Verify proper operation of the solvent extraction system by verifying material balance within the process (The DSS Hold Tank Cs-137 concentration will be very low and the Cs-137 concentration in the Strip Effluent Hold Tank will be fifteen times higher than the Cs-137 concentration in the Feed Tank.) Sodium iodide monitors are used to measure the Cs-137 concentration in the piping before the DSS Hold tank, while GM monitors are used for Cs-137 measurements before the Strip Effluent Hold Tank. Tungsten shields were designed using Monte Carlo calculations and fabricated to reduce the process background radiation at the detector positions. These monitors were calibrated with NIST traceable standards that were specially made to be the same as the piping being monitored. Since this gamma ray monitoring system is unique, specially designed software was written and acceptance tested by Savannah River National Laboratory personnel. The software is a LabView-based application that serves as a unified interface for controlling

  12. 30 CFR 75.336 - Sampling and monitoring requirements.

    Science.gov (United States)

    2010-07-01

    ... concentration of other sampling locations in the sealed area and other required information. Before miners... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Sampling and monitoring requirements. 75.336... SAFETY AND HEALTH MANDATORY SAFETY STANDARDS-UNDERGROUND COAL MINES Ventilation § 75.336 Sampling and...

  13. Specification and Verification of Medical Monitoring System Using Petri-nets.

    Science.gov (United States)

    Majma, Negar; Babamir, Seyed Morteza

    2014-07-01

    To monitor the patient behavior, data are collected from patient's body by a medical monitoring device so as to calculate the output using embedded software. Incorrect calculations may endanger the patient's life if the software fails to meet the patient's requirements. Accordingly, the veracity of the software behavior is a matter of concern in the medicine; moreover, the data collected from the patient's body are fuzzy. Some methods have already dealt with monitoring the medical monitoring devices; however, model based monitoring fuzzy computations of such devices have been addressed less. The present paper aims to present synthesizing a fuzzy Petri-net (FPN) model to verify behavior of a sample medical monitoring device called continuous infusion insulin (INS) because Petri-net (PN) is one of the formal and visual methods to verify the software's behavior. The device is worn by the diabetic patients and then the software calculates the INS dose and makes a decision for injection. The input and output of the infusion INS software are not crisp in the real world; therefore, we present them in fuzzy variables. Afterwards, we use FPN instead of clear PN to model the fuzzy variables. The paper follows three steps to synthesize an FPN to deal with verification of the infusion INS device: (1) Definition of fuzzy variables, (2) definition of fuzzy rules and (3) design of the FPN model to verify the software behavior.

  14. Measuring the electric activity of chick embryos heart through 16 bit audio card monitored by the Goldwavetm software

    Science.gov (United States)

    Silva, Dilson; Cortez, Celia Martins

    2015-12-01

    In the present work we used a high-resolution, low-cost apparatus capable of detecting waves fit inside the sound bandwidth, and the software package GoldwaveTM for graphical display, processing and monitoring the signals, to study aspects of the electric heart activity of early avian embryos, specifically at the 18th Hamburger & Hamilton stage of the embryo development. The species used was the domestic chick (Gallus gallus), and we carried out 23 experiments in which cardiographic spectra of QRS complex waves representing the propagation of depolarization waves through ventricles was recorded using microprobes and reference electrodes directly on the embryos. The results show that technique using 16 bit audio card monitored by the GoldwaveTM software was efficient to study signal aspects of heart electric activity of early avian embryos.

  15. Measurement of software project management effectiveness

    OpenAIRE

    Demir, Kadir Alpaslan

    2008-01-01

    Approved for public release; distribution is unlimited. Evaluating, monitoring, and improving the effectiveness of project management can contribute to successful acquisition of software systems. In this dissertation, we introduce a quantitative metric for gauging the effectiveness of managing a software-development project. The metric may be used to evaluate and monitor project management effectiveness in software projects by project managers, technical managers, executive man...

  16. Interplay between requirements, software architecture, and hardware constraints in the development of a home control user interface

    DEFF Research Database (Denmark)

    Loft, M.S.; Nielsen, S.S.; Nørskov, Kim

    2012-01-01

    is to propose the hardware platform as a third Twin Peaks element that must be given attention in projects such as the one described in this paper. Specifically, we discuss how the presence of severe hardware constraints exacerbates making trade-offs between requirements and architecture.......We have developed a new graphical user interface for a home control device for a large industrial customer. In this industrial case study, we first present our approaches to requirements engineering and to software architecture; we also describe the given hardware platform. Then we make two...... contributions. Our first contribution is to provide a specific example of a real-world project in which a Twin Peaks-compliant approach to software development has been used, and to describe and discuss three examples of interplay between requirements and software architecture decisions. Our second contribution...

  17. Incorporating Gaming in Software Engineering Projects: Case of RMU Monopoly

    Directory of Open Access Journals (Sweden)

    Sushil Acharya

    2009-02-01

    Full Text Available A major challenge in engineering education is retaining student interest in the engineering discipline. Active student involvement in engineering projects is one way of retaining student interest. Such involvement can only be realized if project inception comes entirely from the student. This paper presents a software game, RMU Monopoly, developed as a project requirement for a software engineering course and describes the challenges and gains of implementing such a project. The RMU Monopoly was proposed by three junior software engineering students. The game is a multi-platform software program that allows up to eight players and implements the rules of the Monopoly board game. To ensure agility the game was developed using the spiral software development model. The Software Requirements Specification (SRS document was finalized through an iterative procedure. Standard Unified Modeling Language (UML diagrams were used for product design. A Risk Mitigation, Monitoring, and Management Plan (RMMM was developed to ensure proactive risk management. Gantt chart, weekly progress meetings and weekly scrum meetings were used to track project progress. C# and Sub- Version were used in a client-server architecture to develop the software. The project was successful in retaining student interest in the software engineering discipline

  18. 40 CFR 98.414 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98.414 Section 98.414 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Industrial Greenhouse Gases § 98.414 Monitoring...

  19. Waste Receiving and Processing Facility Module 1 Data Management System software requirements specification

    International Nuclear Information System (INIS)

    Rosnick, C.K.

    1996-01-01

    This document provides the software requirements for Waste Receiving and Processing (WRAP) Module 1 Data Management System (DMS). The DMS is one of the plant computer systems for the new WRAP 1 facility (Project W-0126). The DMS will collect, store and report data required to certify the low level waste (LLW) and transuranic (TRU) waste items processed at WRAP 1 as acceptable for shipment, storage, or disposal

  20. Waste Receiving and Processing Facility Module 1 Data Management System Software Requirements Specification

    International Nuclear Information System (INIS)

    Brann, E.C. II.

    1994-01-01

    This document provides the software requirements for Waste Receiving and Processing (WRAP) Module 1 Data Management System (DMS). The DMS is one of the plant computer systems for the new WRAP 1 facility (Project W-026). The DMS will collect, store and report data required to certify the low level waste (LLW) and transuranic (TRU) waste items processed at WRAP 1 as acceptable for shipment, storage, or disposal

  1. Waste Receiving and Processing Facility Module 1 Data Management System Software Requirements Specification

    Energy Technology Data Exchange (ETDEWEB)

    Brann, E.C. II

    1994-09-09

    This document provides the software requirements for Waste Receiving and Processing (WRAP) Module 1 Data Management System (DMS). The DMS is one of the plant computer systems for the new WRAP 1 facility (Project W-026). The DMS will collect, store and report data required to certify the low level waste (LLW) and transuranic (TRU) waste items processed at WRAP 1 as acceptable for shipment, storage, or disposal.

  2. Deformation integrity monitoring for GNSS positioning services including local, regional and large scale hazard monitoring - the Karlsruhe approach and software(MONIKA)

    Science.gov (United States)

    Jaeger, R.

    2007-05-01

    GNSS-positioning services like SAPOS/ascos in Germany and many others in Europe, America and worldwide, usually yield in a short time their interdisciplinary and country-wide use for precise geo-referencing, replacing traditional low order geodetic networks. So it becomes necessary that possible changes of the reference stations' coordinates are detected ad hoc. The GNSS-reference-station MONitoring by the KArlsruhe approach and software (MONIKA) are designed for that task. The developments at Karlsruhe University of Applied Sciences in cooperation with the State Survey of Baden-Württemberg are further motivated by a the official resolution of the German state survey departments' association (Arbeitsgemeinschaft der Vermessungsverwaltungen Deutschland (AdV)) 2006 on coordinate monitoring as a quality-control duty of the GNSS-positioning service provider. The presented approach can - besides the coordinate control of GNSS-positioning services - also be used to set up any GNSS-service for the tasks of an area-wide geodynamical and natural disaster-prevention service. The mathematical model of approach, which enables a multivariate and multi-epochal design approach, is based on the GNSS-observations input of the RINEX-data of the GNSS service, followed by fully automatic processing of baselines and/or session, and a near-online setting up of epoch-state vectors and their covariance-matrices in a rigorous 3D network adjustment. In case of large scale and long-term monitoring situations, geodynamical standard trends (datum-drift, plate-movements etc.) are accordingly considered and included in the mathematical model of MONIKA. The coordinate-based deformation monitoring approach, as third step of the stepwise adjustments, is based on the above epoch-state vectors, and - splitting off geodynamics trends - hereby on a multivariate and multi-epochal congruency testing. So far, that no other information exists, all points are assumed as being stable and congruent reference

  3. Great software debates

    CERN Document Server

    Davis, A

    2004-01-01

    The industry’s most outspoken and insightful critic explains how the software industry REALLY works. In Great Software Debates, Al Davis, shares what he has learned about the difference between the theory and the realities of business and encourages you to question and think about software engineering in ways that will help you succeed where others fail. In short, provocative essays, Davis fearlessly reveals the truth about process improvement, productivity, software quality, metrics, agile development, requirements documentation, modeling, software marketing and sales, empiricism, start-up financing, software research, requirements triage, software estimation, and entrepreneurship.

  4. Hardware and software constructs for a vibration analysis network

    International Nuclear Information System (INIS)

    Cook, S.A.; Crowe, R.D.; Toffer, H.

    1985-01-01

    Vibration level monitoring and analysis has been initiated at N Reactor, the dual purpose reactor operated at Hanford, Washington by UNC Nuclear Industries (UNC) for the Department of Energy (DOE). The machinery to be monitored was located in several buildings scattered over the plant site, necessitating an approach using satellite stations to collect, monitor and temporarily store data. The satellite stations are, in turn, linked to a centralized processing computer for further analysis. The advantages of a networked data analysis system are discussed in this paper along with the hardware and software required to implement such a system

  5. A Systematic Mapping on Supporting Approaches for Requirements Traceability in the Context of Software Projects

    Directory of Open Access Journals (Sweden)

    MALCHER, P R.C.

    2015-12-01

    Full Text Available The Requirements Traceability is seen as a quality factor with regard to software development, being present in standards and quality models. In this context, several techniques, models, frameworks and tools have been used to support it. Thus, the purpose of this paper is to present a systematic mapping carried out in order to find in the literature approaches to support the requirements traceability in the context of software projects and make the categorization of the data found in order to demonstrate, by means of a reliable, accurate and auditable method, how this area has developed and what are the main approaches are used to implement it.

  6. The Clinical Utilisation of Respiratory Elastance Software (CURE Soft): a bedside software for real-time respiratory mechanics monitoring and mechanical ventilation management.

    Science.gov (United States)

    Szlavecz, Akos; Chiew, Yeong Shiong; Redmond, Daniel; Beatson, Alex; Glassenbury, Daniel; Corbett, Simon; Major, Vincent; Pretty, Christopher; Shaw, Geoffrey M; Benyo, Balazs; Desaive, Thomas; Chase, J Geoffrey

    2014-09-30

    Real-time patient respiratory mechanics estimation can be used to guide mechanical ventilation settings, particularly, positive end-expiratory pressure (PEEP). This work presents a software, Clinical Utilisation of Respiratory Elastance (CURE Soft), using a time-varying respiratory elastance model to offer this ability to aid in mechanical ventilation treatment. CURE Soft is a desktop application developed in JAVA. It has two modes of operation, 1) Online real-time monitoring decision support and, 2) Offline for user education purposes, auditing, or reviewing patient care. The CURE Soft has been tested in mechanically ventilated patients with respiratory failure. The clinical protocol, software testing and use of the data were approved by the New Zealand Southern Regional Ethics Committee. Using CURE Soft, patient's respiratory mechanics response to treatment and clinical protocol were monitored. Results showed that the patient's respiratory elastance (Stiffness) changed with the use of muscle relaxants, and responded differently to ventilator settings. This information can be used to guide mechanical ventilation therapy and titrate optimal ventilator PEEP. CURE Soft enables real-time calculation of model-based respiratory mechanics for mechanically ventilated patients. Results showed that the system is able to provide detailed, previously unavailable information on patient-specific respiratory mechanics and response to therapy in real-time. The additional insight available to clinicians provides the potential for improved decision-making, and thus improved patient care and outcomes.

  7. Qualification of Simulation Software for Safety Assessment of Sodium Cooled Fast Reactors. Requirements and Recommendations

    Energy Technology Data Exchange (ETDEWEB)

    Brown, Nicholas R. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Pointer, William David [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Sieger, Matt [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Flanagan, George F. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Moe, Wayne [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); HolbrookINL, Mark [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2016-04-01

    The goal of this review is to enable application of codes or software packages for safety assessment of advanced sodium-cooled fast reactor (SFR) designs. To address near-term programmatic needs, the authors have focused on two objectives. First, the authors have focused on identification of requirements for software QA that must be satisfied to enable the application of software to future safety analyses. Second, the authors have collected best practices applied by other code development teams to minimize cost and time of initial code qualification activities and to recommend a path to the stated goal.

  8. The whiteStar development project: Westinghouse's next generation core design simulator and core monitoring software to power the nuclear renaissance

    International Nuclear Information System (INIS)

    Boyd, W. A.; Mayhue, L. T.; Penkrot, V. S.; Zhang, B.

    2009-01-01

    The WhiteStar project has undertaken the development of the next generation core analysis and monitoring system for Westinghouse Electric Company. This on-going project focuses on the development of the ANC core simulator, BEACON core monitoring system and NEXUS nuclear data generation system. This system contains many functional upgrades to the ANC core simulator and BEACON core monitoring products as well as the release of the NEXUS family of codes. The NEXUS family of codes is an automated once-through cross section generation system designed for use in both PWR and BWR applications. ANC is a multi-dimensional nodal code for all nuclear core design calculations at a given condition. ANC predicts core reactivity, assembly power, rod power, detector thimble flux, and other relevant core characteristics. BEACON is an advanced core monitoring and support system which uses existing instrumentation data in conjunction with an analytical methodology for on-line generation and evaluation of 3D core power distributions. This new system is needed to design and monitor the Westinghouse AP1000 PWR. This paper describes provides an overview of the software system, software development methodologies used as well some initial results. (authors)

  9. Software development methodology for computer based I&C systems of prototype fast breeder reactor

    International Nuclear Information System (INIS)

    Manimaran, M.; Shanmugam, A.; Parimalam, P.; Murali, N.; Satya Murty, S.A.V.

    2015-01-01

    Highlights: • Software development methodology adopted for computer based I&C systems of PFBR is detailed. • Constraints imposed as part of software requirements and coding phase are elaborated. • Compliance to safety and security requirements are described. • Usage of CASE (Computer Aided Software Engineering) tools during software design, analysis and testing phase are explained. - Abstract: Prototype Fast Breeder Reactor (PFBR) is sodium cooled reactor which is in the advanced stage of construction in Kalpakkam, India. Versa Module Europa bus based Real Time Computer (RTC) systems are deployed for Instrumentation & Control of PFBR. RTC systems have to perform safety functions within the stipulated time which calls for highly dependable software. Hence, well defined software development methodology is adopted for RTC systems starting from the requirement capture phase till the final validation of the software product. V-model is used for software development. IEC 60880 standard and AERB SG D-25 guideline are followed at each phase of software development. Requirements documents and design documents are prepared as per IEEE standards. Defensive programming strategies are followed for software development using C language. Verification and validation (V&V) of documents and software are carried out at each phase by independent V&V committee. Computer aided software engineering tools are used for software modelling, checking for MISRA C compliance and to carry out static and dynamic analysis. Various software metrics such as cyclomatic complexity, nesting depth and comment to code are checked. Test cases are generated using equivalence class partitioning, boundary value analysis and cause and effect graphing techniques. System integration testing is carried out wherein functional and performance requirements of the system are monitored

  10. Software development methodology for computer based I&C systems of prototype fast breeder reactor

    Energy Technology Data Exchange (ETDEWEB)

    Manimaran, M., E-mail: maran@igcar.gov.in; Shanmugam, A.; Parimalam, P.; Murali, N.; Satya Murty, S.A.V.

    2015-10-15

    Highlights: • Software development methodology adopted for computer based I&C systems of PFBR is detailed. • Constraints imposed as part of software requirements and coding phase are elaborated. • Compliance to safety and security requirements are described. • Usage of CASE (Computer Aided Software Engineering) tools during software design, analysis and testing phase are explained. - Abstract: Prototype Fast Breeder Reactor (PFBR) is sodium cooled reactor which is in the advanced stage of construction in Kalpakkam, India. Versa Module Europa bus based Real Time Computer (RTC) systems are deployed for Instrumentation & Control of PFBR. RTC systems have to perform safety functions within the stipulated time which calls for highly dependable software. Hence, well defined software development methodology is adopted for RTC systems starting from the requirement capture phase till the final validation of the software product. V-model is used for software development. IEC 60880 standard and AERB SG D-25 guideline are followed at each phase of software development. Requirements documents and design documents are prepared as per IEEE standards. Defensive programming strategies are followed for software development using C language. Verification and validation (V&V) of documents and software are carried out at each phase by independent V&V committee. Computer aided software engineering tools are used for software modelling, checking for MISRA C compliance and to carry out static and dynamic analysis. Various software metrics such as cyclomatic complexity, nesting depth and comment to code are checked. Test cases are generated using equivalence class partitioning, boundary value analysis and cause and effect graphing techniques. System integration testing is carried out wherein functional and performance requirements of the system are monitored.

  11. Distributed caching mechanism for various MPE software services

    CERN Document Server

    Svec, Andrej

    2017-01-01

    The MPE Software Section provides multiple software services to facilitate the testing and the operation of the CERN Accelerator complex. Continuous growth in the number of users and the amount of processed data result in the requirement of high scalability. Our current priority is to move towards a distributed and properly load balanced set of services based on containers. The aim of this project is to implement the generic caching mechanism applicable to our services and chosen architecture. The project will at first require research about the different aspects of distributed caching (persistence, no gc-caching, cache consistency etc.) and the available technologies followed by the implementation of the chosen solution. In order to validate the correctness and performance of the implementation in the last phase of the project it will be required to implement a monitoring layer and integrate it with the current ELK stack.

  12. Practical requirements for software tools to assist in the validation and verification of hybrid expert systems

    International Nuclear Information System (INIS)

    Singh, G.P.; Cadena, D.; Burgess, J.

    1992-01-01

    Any practical software development effort must remain focused on verification and validation of user requirements. Knowledge-based system development is no different in this regard. In industry today, most expert systems being produced are, in reality, hybrid software systems which, in addition to those components that provide the knowledge base and expert reasoning over the problem domain using various rule-based and object-oriented paradigms, incorporate significant bodies of code based on more traditional software techniques such as database management, graphical user interfaces, hypermedia, spreadsheets, as well as specially developed sequential code. Validation and verification of such hybrid systems must perforce integrate suitable methodologies from all such fields. This paper attempts to provide a broad overview of the practical requirements for methodologies and the concomitant groupware tools which would assist in such an enterprise. These methodologies and groupware tools would facilitate the teamwork efforts necessary to validate and verify all components of such hybrid systems by emphasizing cooperative recording of requirements and negotiated resolutions of any conflicts grounded in a solid understanding of the semantics of such a system

  13. CODAS object monitoring service

    International Nuclear Information System (INIS)

    Wheatley, M.R.; Rainford, M.

    2001-01-01

    The primary Control and Data Acquisition System (CODAS) of JET is based on a TCP/IP network of more than 150 computers. The CODAS computers provide the JET machine control and data acquisition for over 70,000 digital and analog signals. The Object Monitoring Service (OMS) is used by applications for monitoring objects for presentation to the JET machine operators and for the operation of individual software components (such as valve state, access control, mimic definition changes and internal data distribution). Each server typically handles connections from around 60 clients monitoring upwards of 2000 objects. Some servers have over 150 clients and 5000 objects. Acquisition libraries are dynamically linked into a running server as required either to acquire data values for objects or to forward requests to other OMS servers. A mechanism involving dynamic linking allows new libraries to be integrated without stopping or changing running software. OMS provides a very reliable and highly successful 'data-type independent' means of monitoring many different objects. It allows applications to take advantage of new data sources, without the need to change existing code

  14. Report on the working conference on requirements engineering: foundation for software quality (REFSQ'09)

    NARCIS (Netherlands)

    Glinz, Martin; Heymans, Patrick; Persson, Anne; Sindre, Guttorm; Aurum, Aybüke; Madhavji, Nazim; Madhavji, N.; Paech, Barbara; Regev, Gil; Wieringa, Roelf J.

    This report summarizes the presentations and discussions at REFSQ’09, the 15th International Working Conference on Requirements Engineering: Foundation for Software Quality which was held on June 8-9, 2009 in Amsterdam, The Netherlands.

  15. The software and hardware design of a 16 channel online dose rate monitoring system

    International Nuclear Information System (INIS)

    Tang Wenjuan; Yan Yonghong; Yang Shiming; Li Xiaonan; Min Jian

    2011-01-01

    The software and hardware design of a 16 channel online dose rate monitoring system is presented. After being amplified and A/D converted, the output signal of the sensors was sent to a microprocessor through an FPGA, where the low-frequency filter, calculation, temperature compensation and pedestal deduction were accomplished. Such steps corrected the variation of dark current dependent on temperature fluctuations in a effective way, and finally the instantaneous dose rate results with enough precise were obtained. (authors)

  16. Information processing requirements for on-board monitoring of automatic landing

    Science.gov (United States)

    Sorensen, J. A.; Karmarkar, J. S.

    1977-01-01

    A systematic procedure is presented for determining the information processing requirements for on-board monitoring of automatic landing systems. The monitoring system detects landing anomalies through use of appropriate statistical tests. The time-to-correct aircraft perturbations is determined from covariance analyses using a sequence of suitable aircraft/autoland/pilot models. The covariance results are used to establish landing safety and a fault recovery operating envelope via an event outcome tree. This procedure is demonstrated with examples using the NASA Terminal Configured Vehicle (B-737 aircraft). The procedure can also be used to define decision height, assess monitoring implementation requirements, and evaluate alternate autoland configurations.

  17. Monitored Retrievable Storage System Requirements Document. Revision 1

    Energy Technology Data Exchange (ETDEWEB)

    1994-03-01

    This Monitored Retrievable Storage System Requirements Document (MRS-SRD) describes the functions to be performed and technical requirements for a Monitored Retrievable Storage (MRS) facility subelement and the On-Site Transfer and Storage (OSTS) subelement. The MRS facility subelement provides for temporary storage, at a Civilian Radioactive Waste Management System (CRWMS) operated site, of spent nuclear fuel (SNF) contained in an NRC-approved Multi-Purpose Canister (MPC) storage mode, or other NRC-approved storage modes. The OSTS subelement provides for transfer and storage, at Purchaser sites, of spent nuclear fuel (SNF) contained in MPCs. Both the MRS facility subelement and the OSTS subelement are in support of the CRWMS. The purpose of the MRS-SRD is to define the top-level requirements for the development of the MRS facility and the OSTS. These requirements include design, operation, and decommissioning requirements to the extent they impact on the physical development of the MRS facility and the OSTS. The document also presents an overall description of the MRS facility and the OSTS, their functions (derived by extending the functional analysis documented by the Physical System Requirements (PSR) Store Waste Document), their segments, and the requirements allocated to the segments. In addition, the top-level interface requirements of the MRS facility and the OSTS are included. As such, the MRS-SRD provides the technical baseline for the MRS Safety Analysis Report (SAR) design and the OSTS Safety Analysis Report design.

  18. Monitoring Student Activity in Collaborative Software Development

    DEFF Research Database (Denmark)

    Dietsch, Daniel; Podelski, Andreas; Nam, Jaechang

    2013-01-01

    year of studies formed 20 groups and worked collaboratively to develop video games. Throughout the lab, students have to use a variety of tools for managing and developing their projects, such as software version control, static analysis tools, wikis, mailing lists, etc. The students are also supported......This paper presents data analysis from a course on Software Engineering in an effort to identify metrics and techniques that would allow instructor to act proactively and identify patterns of low engagement and inefficient peer collaboration. Over the last two terms, 106 students in their second...... by weekly meetings with teaching assistants and instructors regarding group progress, code quality, and management issues. Through these meetings and their interactions with the software tools, students leave a detailed trace of data related to their individual engagement and their collaboration behavior...

  19. HAZARD ANALYSIS SOFTWARE

    International Nuclear Information System (INIS)

    Sommer, S; Tinh Tran, T.

    2008-01-01

    Washington Safety Management Solutions, LLC developed web-based software to improve the efficiency and consistency of hazard identification and analysis, control selection and classification, and to standardize analysis reporting at Savannah River Site. In the new nuclear age, information technology provides methods to improve the efficiency of the documented safety analysis development process which includes hazard analysis activities. This software provides a web interface that interacts with a relational database to support analysis, record data, and to ensure reporting consistency. A team of subject matter experts participated in a series of meetings to review the associated processes and procedures for requirements and standard practices. Through these meetings, a set of software requirements were developed and compiled into a requirements traceability matrix from which software could be developed. The software was tested to ensure compliance with the requirements. Training was provided to the hazard analysis leads. Hazard analysis teams using the software have verified its operability. The software has been classified as NQA-1, Level D, as it supports the analysis team but does not perform the analysis. The software can be transported to other sites with alternate risk schemes. The software is being used to support the development of 14 hazard analyses. User responses have been positive with a number of suggestions for improvement which are being incorporated as time permits. The software has enforced a uniform implementation of the site procedures. The software has significantly improved the efficiency and standardization of the hazard analysis process

  20. Software Schedules Missions, Aids Project Management

    Science.gov (United States)

    2008-01-01

    NASA missions require advanced planning, scheduling, and management, and the Space Agency has worked extensively to develop the programs and software suites necessary to facilitate these complex missions. These enormously intricate undertakings have hundreds of active components that need constant management and monitoring. It is no surprise, then, that the software developed for these tasks is often applicable in other high-stress, complex environments, like in government or industrial settings. NASA work over the past few years has resulted in a handful of new scheduling, knowledge-management, and research tools developed under contract with one of NASA s partners. These tools have the unique responsibility of supporting NASA missions, but they are also finding uses outside of the Space Program.

  1. Advanced monitoring with complex stream processing

    CERN Multimedia

    CERN. Geneva

    2015-01-01

    Making sense of metrics and logs for service monitoring can be a complicated task. Valuable information is normally scattered across several streams of monitoring data, requiring aggregation, correlation and time-based analysis to promptly detect problems and failures. This presentations shows a solution which is used to support the advanced monitoring of the messaging services provided by the IT Department. It uses Esper, an open-source software product for Complex Event Processing (CEP), that analyses series of events for deriving conclusions from them.

  2. Regulatory requirements and quality assurance of radiation monitoring instruments

    International Nuclear Information System (INIS)

    Narasimharao, K.L.; Sharma, Ranjit

    2005-01-01

    The successful utilisation of radiation sources in the fields of medicine and industry requires the accurate measurement of activity, exposure rate and dose. Many varieties of instruments are in use for measurement of these parameters and new ones are being developed. The criteria for the design of the radiation monitoring instrument include the type and intensity of the radiation, purpose of measurement and ruggedness of the instrument. Quality and reliability of radiation monitoring instruments ensure that individuals are adequately protected. Accuracy, response time and ruggedness are required to be as per the approved/ prescribed guidelines. Regulatory authorities outline the design and performance criteria for radiation monitoring instruments and prescribe the recommendations of international agencies such as IAEA, ICRU and ISO for radiological measurement assurance programme. National Standards Laboratories all over the world prescribe procedures for calibration of various radiation monitoring instruments. The instruments should be calibrated as per these guidelines and should be traceable to national standards. The calibration traceable to national/ international standards and documentation as well as limits stipulated by the competent authority ensures the expected performance of the instrument. (author)

  3. Computer-Aided Software Engineering - An approach to real-time software development

    Science.gov (United States)

    Walker, Carrie K.; Turkovich, John J.

    1989-01-01

    A new software engineering discipline is Computer-Aided Software Engineering (CASE), a technology aimed at automating the software development process. This paper explores the development of CASE technology, particularly in the area of real-time/scientific/engineering software, and a history of CASE is given. The proposed software development environment for the Advanced Launch System (ALS CASE) is described as an example of an advanced software development system for real-time/scientific/engineering (RT/SE) software. The Automated Programming Subsystem of ALS CASE automatically generates executable code and corresponding documentation from a suitably formatted specification of the software requirements. Software requirements are interactively specified in the form of engineering block diagrams. Several demonstrations of the Automated Programming Subsystem are discussed.

  4. A Framework for the Management of Evolving Requirements in Software Systems Supporting Network-Centric Warfare

    National Research Council Canada - National Science Library

    Reynolds, Linda K

    2006-01-01

    .... There are many sources of requirements for these software systems supporting NCO, which may increase in number as the Services continue to develop the capabilities necessary for the transformation...

  5. A GIS Software Toolkit for Monitoring Areal Snow Cover and Producing Daily Hydrologic Forecasts using NASA Satellite Imagery, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — Aniuk Consulting, LLC, proposes to create a GIS software toolkit for monitoring areal snow cover extent and producing streamflow forecasts. This toolkit will be...

  6. Diversity requirements for safety critical software-based automation systems

    International Nuclear Information System (INIS)

    Korhonen, J.; Pulkkinen, U.; Haapanen, P.

    1998-03-01

    System vendors nowadays propose software-based systems even for the most critical safety functions in nuclear power plants. Due to the nature and mechanisms of influence of software faults new methods are needed for the safety and reliability evaluation of these systems. In the research project 'Programmable automation systems in nuclear power plants (OHA)' various safety assessment methods and tools for software based systems are developed and evaluated. This report first discusses the (common cause) failure mechanisms in software-based systems, then defines fault-tolerant system architectures to avoid common cause failures, then studies the various alternatives to apply diversity and their influence on system reliability. Finally, a method for the assessment of diversity is described. Other recently published reports in OHA-report series handles the statistical reliability assessment of software based (STUK-YTO-TR 119), usage models in reliability assessment of software-based systems (STUK-YTO-TR 128) and handling of programmable automation in plant PSA-studies (STUK-YTO-TR 129)

  7. Software quality management

    International Nuclear Information System (INIS)

    Bishop, D.C.; Pymm, P.

    1991-01-01

    As programmable electronic (software-based) systems are increasingly being proposed as design solutions for high integrity applications in nuclear power stations, the need to adopt suitable quality management arrangements is paramount. The authors describe Scottish Nuclear's strategy for software quality management and, using the main on-line monitoring system at Torness Power Station as an example, explain how this strategy is put into practice. Particular attention is given to the topics of software quality planning and change control. (author)

  8. Importance of Requirements Analysis & Traceability to Improve Software Quality and Reduce Cost and Risk

    Science.gov (United States)

    Kapoor, Manju M.; Mehta, Manju

    2010-01-01

    The goal of this paper is to emphasize the importance of developing complete and unambiguous requirements early in the project cycle (prior to Preliminary Design Phase). Having a complete set of requirements early in the project cycle allows sufficient time to generate a traceability matrix. Requirements traceability and analysis are the key elements in improving verification and validation process, and thus overall software quality. Traceability can be most beneficial when the system changes. If changes are made to high-level requirements it implies that low-level requirements need to be modified. Traceability ensures that requirements are appropriately and efficiently verified at various levels whereas analysis ensures that a rightly interpreted set of requirements is produced.

  9. Software engineering the current practice

    CERN Document Server

    Rajlich, Vaclav

    2011-01-01

    INTRODUCTION History of Software EngineeringSoftware PropertiesOrigins of SoftwareBirth of Software EngineeringThird Paradigm: Iterative ApproachSoftware Life Span ModelsStaged ModelVariants of Staged ModelSoftware Technologies Programming Languages and CompilersObject-Oriented TechnologyVersion Control SystemSoftware ModelsClass DiagramsUML Activity DiagramsClass Dependency Graphs and ContractsSOFTWARE CHANGEIntroduction to Software ChangeCharacteristics of Software ChangePhases of Software ChangeRequirements and Their ElicitationRequirements Analysis and Change InitiationConcepts and Concept

  10. The tritium monitoring requirements of fusion and the status of research

    International Nuclear Information System (INIS)

    Nickerson, S.B.; Gerdingh, R.F.; Penfold, K.

    1982-10-01

    This report is a summary of an investigation into the tritium monitoring requirements of tritium laboratories, D-T burning ignition experiments, and fusion reactors. There is also a summary of the status of research into tritium monitoring and a survey of commercially available tritium monitors

  11. Synchronized analysis of testbeam data with the Judith software

    CERN Document Server

    McGoldrick, Garrin; Gorišek, Andrej

    2014-01-01

    The Judith software performs pixel detector analysis tasks utilizing two different data streams such as those produced by the reference and tested devices typically found in a testbeam. This software addresses and fixes problems arising from the desynchronization of the two simultaneously triggered data streams by detecting missed triggers in either of the streams. The software can perform all tasks required to generate particle tracks using multiple detector planes: it can align the planes, cluster hits and generate tracks from these clusters. This information can then be used to measure the properties of a particle detector with very fine spatial resolution. It was tested at DESY in the Kartel telescope, a silicon tracking detector, with ATLAS Diamond Beam Monitor modules as a device under test.

  12. Software Engineering Improvement Plan

    Science.gov (United States)

    2006-01-01

    In performance of this task order, bd Systems personnel provided support to the Flight Software Branch and the Software Working Group through multiple tasks related to software engineering improvement and to activities of the independent Technical Authority (iTA) Discipline Technical Warrant Holder (DTWH) for software engineering. To ensure that the products, comments, and recommendations complied with customer requirements and the statement of work, bd Systems personnel maintained close coordination with the customer. These personnel performed work in areas such as update of agency requirements and directives database, software effort estimation, software problem reports, a web-based process asset library, miscellaneous documentation review, software system requirements, issue tracking software survey, systems engineering NPR, and project-related reviews. This report contains a summary of the work performed and the accomplishments in each of these areas.

  13. Research of real-time communication software

    Science.gov (United States)

    Li, Maotang; Guo, Jingbo; Liu, Yuzhong; Li, Jiahong

    2003-11-01

    Real-time communication has been playing an increasingly important role in our work, life and ocean monitor. With the rapid progress of computer and communication technique as well as the miniaturization of communication system, it is needed to develop the adaptable and reliable real-time communication software in the ocean monitor system. This paper involves the real-time communication software research based on the point-to-point satellite intercommunication system. The object-oriented design method is adopted, which can transmit and receive video data and audio data as well as engineering data by satellite channel. In the real-time communication software, some software modules are developed, which can realize the point-to-point satellite intercommunication in the ocean monitor system. There are three advantages for the real-time communication software. One is that the real-time communication software increases the reliability of the point-to-point satellite intercommunication system working. Second is that some optional parameters are intercalated, which greatly increases the flexibility of the system working. Third is that some hardware is substituted by the real-time communication software, which not only decrease the expense of the system and promotes the miniaturization of communication system, but also aggrandizes the agility of the system.

  14. Integrated environmental monitoring -- prototype demonstration

    International Nuclear Information System (INIS)

    Bryce, R.W.; Vail, L.W.; Hostetler, D.D.; Meyer, P.D.; Carlson, T.J.; Miller, P.L.

    1994-01-01

    Groundwater monitoring is an important activity at US Department of Energy (DOE) sites. Monitoring programs at DOE facilities have evolved in response to operational needs at the facilities, public outcries for information, regulatory requirements, DOE orders, and improvements in monitoring technology. Decisions regarding sampling location, sampling frequency, analyses performed, and other aspects of monitoring network design can have major implications for detecting releases and for making subsequent higher level decisions about facility operation and remediation. The Integrated Environmental Monitoring (IEM) concept is a set of analytical procedures and software tools that can be used to improve monitoring network design decisions. Such decisions include the choice of monitoring locations, sampling frequencies, sensor technologies, and monitored constituents. IEM provides a set of monitoring alternatives that balance the tradeoffs between competing monitoring objectives such as the minimization of cost and the minimization of uncertainty. The alternatives provided are the best available with respect to the monitoring objectives, consistent with the physical and chemical characteristics of the site, and consist with applicable regulatory requirements. The selection of the best monitoring alternative to implement is made by the stakeholders after reviewing the alternatives and tradeoffs produced by the IEM process

  15. Software quality assurance: in large scale and complex software-intensive systems

    NARCIS (Netherlands)

    Mistrik, I.; Soley, R.; Ali, N.; Grundy, J.; Tekinerdogan, B.

    2015-01-01

    Software Quality Assurance in Large Scale and Complex Software-intensive Systems presents novel and high-quality research related approaches that relate the quality of software architecture to system requirements, system architecture and enterprise-architecture, or software testing. Modern software

  16. Evaluation of a Game to Teach Requirements Collection and Analysis in Software Engineering at Tertiary Education Level

    Science.gov (United States)

    Hainey, Thomas; Connolly, Thomas M.; Stansfield, Mark; Boyle, Elizabeth A.

    2011-01-01

    A highly important part of software engineering education is requirements collection and analysis which is one of the initial stages of the Database Application Lifecycle and arguably the most important stage of the Software Development Lifecycle. No other conceptual work is as difficult to rectify at a later stage or as damaging to the overall…

  17. Software Defined Networking Demands on Software Technologies

    DEFF Research Database (Denmark)

    Galinac Grbac, T.; Caba, Cosmin Marius; Soler, José

    2015-01-01

    Software Defined Networking (SDN) is a networking approach based on a centralized control plane architecture with standardised interfaces between control and data planes. SDN enables fast configuration and reconfiguration of the network to enhance resource utilization and service performances....... This new approach enables a more dynamic and flexible network, which may adapt to user needs and application requirements. To this end, systemized solutions must be implemented in network software, aiming to provide secure network services that meet the required service performance levels. In this paper......, we review this new approach to networking from an architectural point of view, and identify and discuss some critical quality issues that require new developments in software technologies. These issues we discuss along with use case scenarios. Here in this paper we aim to identify challenges...

  18. From Software Development to Software Assembly

    NARCIS (Netherlands)

    Sneed, Harry M.; Verhoef, Chris

    2016-01-01

    The lack of skilled programming personnel and the growing burden of maintaining customized software are forcing organizations to quit producing their own software. It's high time they turned to ready-made, standard components to fulfill their business requirements. Cloud services might be one way to

  19. Hardware and software improvements to a low-cost horizontal parallax holographic video monitor.

    Science.gov (United States)

    Henrie, Andrew; Codling, Jesse R; Gneiting, Scott; Christensen, Justin B; Awerkamp, Parker; Burdette, Mark J; Smalley, Daniel E

    2018-01-01

    Displays capable of true holographic video have been prohibitively expensive and difficult to build. With this paper, we present a suite of modularized hardware components and software tools needed to build a HoloMonitor with basic "hacker-space" equipment, highlighting improvements that have enabled the total materials cost to fall to $820, well below that of other holographic displays. It is our hope that the current level of simplicity, development, design flexibility, and documentation will enable the lay engineer, programmer, and scientist to relatively easily replicate, modify, and build upon our designs, bringing true holographic video to the masses.

  20. Global Software Engineering: A Software Process Approach

    Science.gov (United States)

    Richardson, Ita; Casey, Valentine; Burton, John; McCaffery, Fergal

    Our research has shown that many companies are struggling with the successful implementation of global software engineering, due to temporal, cultural and geographical distance, which causes a range of factors to come into play. For example, cultural, project managementproject management and communication difficulties continually cause problems for software engineers and project managers. While the implementation of efficient software processes can be used to improve the quality of the software product, published software process models do not cater explicitly for the recent growth in global software engineering. Our thesis is that global software engineering factors should be included in software process models to ensure their continued usefulness in global organisations. Based on extensive global software engineering research, we have developed a software process, Global Teaming, which includes specific practices and sub-practices. The purpose is to ensure that requirements for successful global software engineering are stipulated so that organisations can ensure successful implementation of global software engineering.

  1. Intelligent Hardware-Enabled Sensor and Software Safety and Health Management for Autonomous UAS

    Science.gov (United States)

    Rozier, Kristin Y.; Schumann, Johann; Ippolito, Corey

    2015-01-01

    Unmanned Aerial Systems (UAS) can only be deployed if they can effectively complete their mission and respond to failures and uncertain environmental conditions while maintaining safety with respect to other aircraft as well as humans and property on the ground. We propose to design a real-time, onboard system health management (SHM) capability to continuously monitor essential system components such as sensors, software, and hardware systems for detection and diagnosis of failures and violations of safety or performance rules during the ight of a UAS. Our approach to SHM is three-pronged, providing: (1) real-time monitoring of sensor and software signals; (2) signal analysis, preprocessing, and advanced on-the- y temporal and Bayesian probabilistic fault diagnosis; (3) an unobtrusive, lightweight, read-only, low-power hardware realization using Field Programmable Gate Arrays (FPGAs) in order to avoid overburdening limited computing resources or costly re-certi cation of ight software due to instrumentation. No currently available SHM capabilities (or combinations of currently existing SHM capabilities) come anywhere close to satisfying these three criteria yet NASA will require such intelligent, hardwareenabled sensor and software safety and health management for introducing autonomous UAS into the National Airspace System (NAS). We propose a novel approach of creating modular building blocks for combining responsive runtime monitoring of temporal logic system safety requirements with model-based diagnosis and Bayesian network-based probabilistic analysis. Our proposed research program includes both developing this novel approach and demonstrating its capabilities using the NASA Swift UAS as a demonstration platform.

  2. 40 CFR 141.87 - Monitoring requirements for water quality parameters.

    Science.gov (United States)

    2010-07-01

    ... § 141.87 Monitoring requirements for water quality parameters. All large water systems, and all small- and medium-size systems that exceed the lead or copper action level shall monitor water quality... methods. (i) Tap samples shall be representative of water quality throughout the distribution system...

  3. RELAP-7 Software Verification and Validation Plan: Requirements Traceability Matrix (RTM) Part 1 – Physics and numerical methods

    Energy Technology Data Exchange (ETDEWEB)

    Choi, Yong Joon [Idaho National Lab. (INL), Idaho Falls, ID (United States); Yoo, Jun Soo [Idaho National Lab. (INL), Idaho Falls, ID (United States); Smith, Curtis Lee [Idaho National Lab. (INL), Idaho Falls, ID (United States)

    2015-09-01

    This INL plan comprehensively describes the Requirements Traceability Matrix (RTM) on main physics and numerical method of the RELAP-7. The plan also describes the testing-based software verification and validation (SV&V) process—a set of specially designed software models used to test RELAP-7.

  4. Requirements Report Computer Software System for a Semi-Automatic Pipe Handling System and Fabrication Facility

    National Research Council Canada - National Science Library

    1980-01-01

    .... This report is to present the requirements of the computer software that must be developed to create Pipe Detail Drawings and to support the processing of the Pipe Detail Drawings through the Pipe Shop...

  5. Software Atom: An approach towards software components structuring to improve reusability

    Directory of Open Access Journals (Sweden)

    Muhammad Hussain Mughal

    2017-12-01

    Full Text Available Diversity of application domain compelled to design sustainable classification scheme for significantly amassing software repository. The atomic reusable software components are articulated to improve the software component reusability in volatile industry.  Numerous approaches of software classification have been proposed over past decades. Each approach has some limitations related to coupling and cohesion. In this paper, we proposed a novel approach by constituting the software based on radical functionalities to improve software reusability. We analyze the element's semantics in Periodic Table used in chemistry to design our classification approach, and present this approach using tree-based classification to curtail software repository search space complexity and further refined based on semantic search techniques. We developed a Global unique Identifier (GUID for indexing the functions and related components. We have exploited the correlation between chemistry element and software elements to simulate one to one mapping between them. Our approach is inspired from sustainability chemical periodic table. We have proposed software periodic table (SPT representing atomic software components extracted from real application software. Based on SPT classified repository tree parsing & extraction to enable the user to program their software by customizing the ingredients of software requirements. The classified repository of software ingredients assist user to exploits their requirements to software engineer and enable requirement engineer to develop a rapid large-scale prototype with great essence. Furthermore, we would predict the usability of the categorized repository based on feedback of users.  The continuous evolution of that proposed repository will be fine-tuned based on utilization and SPT would be gradually optimized by ant colony optimization techniques. Succinctly would provoke automating the software development process.

  6. A concept of software testing for SMART MMIS software

    International Nuclear Information System (INIS)

    Seo, Yong Seok; Seong, Seung Hwan; Park, Keun Ok; Hur, Sub; Kim, Dong Hoon

    2001-01-01

    In order to achieve high quality of SMART MMIS software, the well-constructed software testing concept shall be required. This paper established software testing concept which is to be applied to SMART MMIS software, in terms of software testing organization, documentation. procedure, and methods. The software testing methods are classified into source code static analysis and dynamic testing. The software dynamic testing methods are discussed with two aspects: white-box and black-box testing. As software testing concept introduced in this paper is applied to the SMART MMIS software. the high quality of the software will be produced. In the future, software failure data will be collected through the construction of SMART MMIS prototyping facility which the software testing concept of this paper is applied to

  7. Review of monitoring uncertainty requirements in the CDM

    International Nuclear Information System (INIS)

    Shishlov, Igor; Bellassen, Valentin

    2014-10-01

    In order to ensure the environmental integrity of carbon offset projects, emission reductions certified under the Clean Development Mechanism (CDM) have to be 'real, measurable and additional', which is ensured through the monitoring, reporting and verification (MRV) process. MRV, however, comes at a cost that ranges from several cents to EUR1.20 and above per ton of CO 2 e depending on the project type. This article analyzes monitoring uncertainty requirements for carbon offset projects with a particular focus on the trade-off between monitoring stringency and cost. To this end, we review existing literature, scrutinize both overarching monitoring guidelines and the 10 most-used methodologies, and finally we analyze four case studies. We find that there is indeed a natural trade-off between the stringency and the cost of monitoring, which if not addressed properly may become a major barrier for the implementation of offset projects in some sectors. We demonstrate that this trade-off has not been systematically addressed in the overarching CDM guidelines and that there are only limited incentives to reduce monitoring uncertainty. Some methodologies and calculation tools as well as some other offset standards, however, do incorporate provisions for a trade-off between monitoring costs and stringency. These provisions may take the form of discounting emissions reductions based on the level of monitoring uncertainty - or more implicitly through allowing a project developer to choose between monitoring a given parameter and using a conservative default value. Our findings support the introduction of an uncertainty standard under the CDM for more comprehensive, yet cost-efficient, accounting for monitoring uncertainty in carbon offset projects. (authors)

  8. InDEx: Open Source iOS and Android Software for Self-Reporting and Monitoring of Alcohol Consumption.

    Science.gov (United States)

    Leightley, Daniel; Puddephatt, Jo-Anne; Goodwin, Laura; Rona, Roberto; Fear, Nicola T

    2018-03-23

    InDEx is a software package for reporting and monitoring alcohol consumption via a smartphone application. Consumption of alcohol is self-reported by the user, and the app provides a visual representation of drinking behaviour and offers feedback on consumption levels compared to the general population. InDEx is intended as an exemplar app, operating as a standalone smartphone application and is highly customisable for a variety of research domains. InDEx is written in JavaScript, using IONIC framework which is cross-platform and is available under the liberal GNU General Public License (v3). The software is available from GitHub (https://github.com/DrDanL/index-app-public).

  9. Power consumption monitoring using additional monitoring device

    Energy Technology Data Exchange (ETDEWEB)

    Truşcă, M. R. C., E-mail: radu.trusca@itim-cj.ro; Albert, Ş., E-mail: radu.trusca@itim-cj.ro; Tudoran, C., E-mail: radu.trusca@itim-cj.ro; Soran, M. L., E-mail: radu.trusca@itim-cj.ro; Fărcaş, F., E-mail: radu.trusca@itim-cj.ro [National Institute for Research and Development of Isotopic and Molecular Technologies, 65-103 Donath, 400293 Cluj-Napoca (Romania); Abrudean, M. [Technical University of Cluj-Napoca, Cluj-Napoca (Romania)

    2013-11-13

    Today, emphasis is placed on reducing power consumption. Computers are large consumers; therefore it is important to know the total consumption of computing systems. Since their optimal functioning requires quite strict environmental conditions, without much variation in temperature and humidity, reducing energy consumption cannot be made without monitoring environmental parameters. Thus, the present work uses a multifunctional electric meter UPT 210 for power consumption monitoring. Two applications were developed: software which carries meter readings provided by electronic and programming facilitates remote device and a device for temperature monitoring and control. Following temperature variations that occur both in the cooling system, as well as the ambient, can reduce energy consumption. For this purpose, some air conditioning units or some computers are stopped in different time slots. These intervals were set so that the economy is high, but the work's Datacenter is not disturbed.

  10. Development and application of all-digital monitoring system

    International Nuclear Information System (INIS)

    Xu Tao; Li Jing; Wang Wei

    2014-01-01

    All digital control system has developed into a mainstream means of monitoring, and achieved information, intelligence, and networking. All-digital control system is characterized by clear image, large transport stream, so the higher the data storage and network bandwidth should be required. Existing analog surveillance system architecture, hardware and software configuration can not meet the requirements of all-digital monitoring system, so how to solve the original analog surveillance system is gradually transformed into fully digital monitoring system, to avoid incompatibility issues in surveillance monitoring system upgrade become a research project. This paper describes the advantages and future direction of megapixels camera and proposes key technologies to solve the resolution and frame rate with the actual project requirements, achieves a core technology of megapixels video surveillance system, and proposes solutions for the actual renovation project problems. (authors)

  11. IDENTIFICATION OF DOE'S POST-CLOSURE MONITORING NEEDS AND REQUIREMENTS

    Energy Technology Data Exchange (ETDEWEB)

    M.A. Ebadian, Ph.D.

    1999-01-01

    The 2006 plan sets an ambitious agenda for the U.S. Department of Energy (DOE), Office of Environmental Management (EM) and the remediation of sites contaminated by decades of nuclear weapons production activities. The plan's primary objective is to reduce overall clean up costs by first eliminating the environmental problems that are most expensive to control and safely maintain. In the context of the 2006 Plan, closure refers to the completion of area or facility specific cleanup projects. The cleanup levels are determined by the planned future use of the site or facility. Use restrictions are still undecided for most sites but are highly probable to exclude residential or agricultural activities. Most of the land will be remediated to ''industrial use'' levels with access restrictions and some areas will be closed-off through containment. Portions of the site will be reserved for waste disposal, either as a waste repository or the in-situ immobilization of contaminated soil and groundwater, and land use will be restricted to waste disposal only. The land used for waste disposal will require monitoring and maintenance activities after closure. Most of the land used for industrial use may also require such postclosure activities. The required postclosure monitoring and maintenance activities will be imposed by regulators and stakeholders. Regulators will not approve closure plans without clearly defined monitoring methods using approved technologies. Therefore, among all other more costly and labor-intensive closure-related activities, inadequate planning for monitoring and lack of appropriate monitoring technologies can prevent closure. The purpose of this project is to determine, document, and track the current and evolving postclosure monitoring requirements at DOE-EM sites. This information will aid CMST-CP in guiding its postclosure technology development and deployment efforts.

  12. Clinical evaluation of monitor unit software and the application of action levels

    International Nuclear Information System (INIS)

    Georg, Dietmar; Nyholm, Tufve; Olofsson, Joergen; Kjaer-Kristoffersen, Flemming; Schnekenburger, Bruno; Winkler, Peter; Nystroem, Hakan; Ahnesjoe, Anders; Karlsson, Mikael

    2007-01-01

    Purpose: The aim of this study was the clinical evaluation of an independent dose and monitor unit verification (MUV) software which is based on sophisticated semi-analytical modelling. The software was developed within the framework of an ESTRO project. Finally, consistent handling of dose calculation deviations applying individual action levels is discussed. Materials and methods: A Matlab-based software ('MUV') was distributed to five well-established treatment centres in Europe (Vienna, Graz, Basel, Copenhagen, and Umea) and evaluated as a quality assurance (QA) tool in clinical routine. Results were acquired for 226 individual treatment plans including a total of 815 radiation fields. About 150 beam verification measurements were performed for a portion of the individual treatment plans, mainly with time variable fluence patterns. The deviations between dose calculations performed with a treatment planning system (TPS) and the MUV software were scored with respect to treatment area, treatment technique, geometrical depth, radiological depth, etc. Results: In general good agreement was found between calculations performed with the different TPSs and MUV, with a mean deviation per field of 0.2 ± 3.5% (1 SD) and mean deviations of 0.2 ± 2.2% for composite treatment plans. For pelvic treatments less than 10% of all fields showed deviations larger than 3%. In general, when using the radiological depth for verification calculations the results and the spread in the results improved significantly, especially for head-and-neck and for thorax treatments. For IMRT head-and-neck beams, mean deviations between MUV and the local TPS were -1.0 ± 7.3% for dynamic, and -1.3 ± 3.2% for step-and-shoot IMRT delivery. For dynamic IMRT beams in the pelvis good agreement was obtained between MUV and the local TPS (mean: -1.6 ± 1.5%). Treatment site and treatment technique dependent action levels between ±3% and ±5% seem to be clinically realistic if a radiological depth

  13. A Business Analytics Software Tool for Monitoring and Predicting Radiology Throughput Performance.

    Science.gov (United States)

    Jones, Stephen; Cournane, Seán; Sheehy, Niall; Hederman, Lucy

    2016-12-01

    Business analytics (BA) is increasingly being utilised by radiology departments to analyse and present data. It encompasses statistical analysis, forecasting and predictive modelling and is used as an umbrella term for decision support and business intelligence systems. The primary aim of this study was to determine whether utilising BA technologies could contribute towards improved decision support and resource management within radiology departments. A set of information technology requirements were identified with key stakeholders, and a prototype BA software tool was designed, developed and implemented. A qualitative evaluation of the tool was carried out through a series of semi-structured interviews with key stakeholders. Feedback was collated, and emergent themes were identified. The results indicated that BA software applications can provide visibility of radiology performance data across all time horizons. The study demonstrated that the tool could potentially assist with improving operational efficiencies and management of radiology resources.

  14. Microbiological Monitoring for the Constellation Program: Current Requirements and Future Considerations

    Science.gov (United States)

    Ott, C. Mark

    2007-01-01

    Microbiological requirements for spaceflight are based on assessments of infectious disease risk which could impact crew health or mission success. The determination of risk from infectious disease is composed of several factors including (1) crew susceptibility, (2) crew exposure to the infectious disease agent, (3) the concentration of the infectious agent, and (4) the characteristics of the infectious agent. As a result of the Health Stabilization Program, stringent monitoring, and cleaning protocols, in-flight environmental microbial monitoring is not necessary for short-duration spaceflights. However, risk factors change for long-duration missions, as exemplified by the presence of medically significant organisms in the environments of both the Mir and International Space Station (ISS). Based upon this historical evidence, requirements for short duration usage aboard the Orion Crew Exploration Vehicle and Lunar Lander Vehicle will not require in-flight monitoring; however, as mission duration increases with a Lunar Outpost, an ability to detect microbial hazard will be necessary. The nature of the detection requirements will depend on the maturity of technology in a rapidly evolving marketplace. Regardless, the hardware will still need to maximize information to discipline experts and the crew, while minimizing the size, mass, power consumption, and crew time usage. The refinement of these monitors will be a major goal in our efforts to travel successfully to Mars.

  15. Software Radar Technology

    Directory of Open Access Journals (Sweden)

    Tang Jun

    2015-08-01

    Full Text Available In this paper, the definition and the key features of Software Radar, which is a new concept, are proposed and discussed. We consider the development of modern radar system technology to be divided into three stages: Digital Radar, Software radar and Intelligent Radar, and the second stage is just commencing now. A Software Radar system should be a combination of various modern digital modular components conformed to certain software and hardware standards. Moreover, a software radar system with an open system architecture supporting to decouple application software and low level hardware would be easy to adopt "user requirements-oriented" developing methodology instead of traditional "specific function-oriented" developing methodology. Compared with traditional Digital Radar, Software Radar system can be easily reconfigured and scaled up or down to adapt to the changes of requirements and technologies. A demonstration Software Radar signal processing system, RadarLab 2.0, which has been developed by Tsinghua University, is introduced in this paper and the suggestions for the future development of Software Radar in China are also given in the conclusion.

  16. Autonomous smart sensor network for full-scale structural health monitoring

    Science.gov (United States)

    Rice, Jennifer A.; Mechitov, Kirill A.; Spencer, B. F., Jr.; Agha, Gul A.

    2010-04-01

    The demands of aging infrastructure require effective methods for structural monitoring and maintenance. Wireless smart sensor networks offer the ability to enhance structural health monitoring (SHM) practices through the utilization of onboard computation to achieve distributed data management. Such an approach is scalable to the large number of sensor nodes required for high-fidelity modal analysis and damage detection. While smart sensor technology is not new, the number of full-scale SHM applications has been limited. This slow progress is due, in part, to the complex network management issues that arise when moving from a laboratory setting to a full-scale monitoring implementation. This paper presents flexible network management software that enables continuous and autonomous operation of wireless smart sensor networks for full-scale SHM applications. The software components combine sleep/wake cycling for enhanced power management with threshold detection for triggering network wide tasks, such as synchronized sensing or decentralized modal analysis, during periods of critical structural response.

  17. Software factory techniques applied to process control at CERN

    OpenAIRE

    Dutour, Mathias D

    2007-01-01

    The CERN Large Hadron Collider (LHC) requires constant monitoring and control of quantities of parameters to guarantee operational conditions. For this purpose, a methodology called UNICOS (UNIfied Industrial COntrols Systems) has been implemented to standardize the design of process control applications. To further accelerate the development of these applications, we migrated our existing UNICOS tooling suite toward a software factory in charge of assembling project, domain and technical inf...

  18. Software engineering for the EBR-II data acquisition system conversion

    International Nuclear Information System (INIS)

    Schorzman, W.

    1988-01-01

    The original data acquisition system (DAS) for the Experimental Breeder Reactor II (EBR-II) was placed into service with state-of-the-art computer and peripherals in 1970. Software engineering principles for real-time data acquisition were in their infancy, and the original software design was dictated by limited hardware resources. The functional requirements evolved from creative ways to gather and display data. This abstract concept developed into an invaluable tool for system analysis, data reporting, and as a plant monitor for operations. In this paper the approach is outlined to the software conversion project with the restraints of operational transparency and 6 weeks for final conversion and testing. The outline is then compared with the formal principles of software engineering to show the way that bridge the gap can be bridged between the theoretical and real world by analyzing the work and listing the lessons learned

  19. Software testability and its application to avionic software

    Science.gov (United States)

    Voas, Jeffrey M.; Miller, Keith W.; Payne, Jeffery E.

    1993-01-01

    Randomly generated black-box testing is an established yet controversial method of estimating software reliability. Unfortunately, as software applications have required higher reliabilities, practical difficulties with black-box testing have become increasingly problematic. These practical problems are particularly acute in life-critical avionics software, where requirements of 10 exp -7 failures per hour of system reliability can translate into a probability of failure (POF) of perhaps 10 exp -9 or less for each individual execution of the software. This paper describes the application of one type of testability analysis called 'sensitivity analysis' to B-737 avionics software; one application of sensitivity analysis is to quantify whether software testing is capable of detecting faults in a particular program and thus whether we can be confident that a tested program is not hiding faults. We so 80 by finding the testabilities of the individual statements of the program, and then use those statement testabilities to find the testabilities of the functions and modules. For the B-737 system we analyzed, we were able to isolate those functions that are more prone to hide errors during system/reliability testing.

  20. [Software as medical devices/medical apps : Tasks, requirements, and experiences from the point of view of a competent authority].

    Science.gov (United States)

    Terhechte, Arno

    2018-03-01

    Software can be classified as a medical device according to the Medical Device Directive 93/42/EEC. The number of software products and medical apps is continuously increasing and so too is the use in health institutions (e. g., in hospitals and doctors' surgeries) for diagnosis and therapy.Different aspects of standalone software and medical apps from the perspective of the authority responsible are presented. The quality system implemented to establish a risk-based systematic inspection and supervision of manufacturers is discussed. The legal framework, as well as additional standards that are the basis for inspection, are outlined. The article highlights special aspects that occur during inspection like verification of software and interfaces, and the clinical evaluation of software. The Bezirksregierung, as the local government authority responsible in North Rhine-Westphalia, is also in charge of inspection of health institutions. Therefore this article is not limited to the manufacturers placing the software on the market, but in addition it describes the management and use of software as a medical device in hospitals.The future legal framework, the Medical Device Regulation, will strengthen the requirements and engage notified bodies more than today in the conformity assessment of software as a medical device.Manufacturers, health institutions, notified bodies and the authorities responsible are in charge of intensifying their efforts towards software as a medical device. Mutual information, improvement of skills, and inspections will lead to compliance with regulatory requirements.

  1. Software factory techniques applied to process control at CERN

    CERN Document Server

    Dutour, Mathias D

    2008-01-01

    The CERN Large Hadron Collider (LHC) requires constant monitoring and control of quantities of parameters to guarantee operational conditions. For this purpose, a methodology called UNICOS (UNIfied Industrial COntrols Systems) has been implemented to standardize the design of process control applications. To further accelerate the development of these applications, we migrated our existing UNICOS tooling suite toward a software factory in charge of assembling project, domain and technical information seamlessly into deployable PLC (Programmable logic Controller) - SCADA (Supervisory Control And Data Acquisition) systems. This software factory delivers consistently high quality by reducing human error and repetitive tasks, and adapts to user specifications in a cost-efficient way. Hence, this production tool is designed to encapsulate and hide the PLC and SCADA target platforms, enabling the experts to focus on the business model rather than specific syntaxes and grammars. Based on industry standard software, ...

  2. The Software Invention Cube: A classification scheme for software inventions

    NARCIS (Netherlands)

    Bergstra, J.A.; Klint, P.

    2008-01-01

    The patent system protects inventions. The requirement that a software invention should make ‘a technical contribution’ turns out to be untenable in practice and this raises the question, what constitutes an invention in the realm of software. The authors developed the Software Invention Cube

  3. An industrial case study in reconstructing requirements views

    NARCIS (Netherlands)

    Lormans, M.; Van Deursen, A.; Gross, H.G.

    2008-01-01

    Requirements views, such as coverage and status views, are an important asset for monitoring and managing software development projects. We have developed a method that automates the process of reconstructing these views, and we have built a tool, REQANALYST, that supports this method. This paper

  4. Software Reliability Issues Concerning Large and Safety Critical Software Systems

    Science.gov (United States)

    Kamel, Khaled; Brown, Barbara

    1996-01-01

    This research was undertaken to provide NASA with a survey of state-of-the-art techniques using in industrial and academia to provide safe, reliable, and maintainable software to drive large systems. Such systems must match the complexity and strict safety requirements of NASA's shuttle system. In particular, the Launch Processing System (LPS) is being considered for replacement. The LPS is responsible for monitoring and commanding the shuttle during test, repair, and launch phases. NASA built this system in the 1970's using mostly hardware techniques to provide for increased reliability, but it did so often using custom-built equipment, which has not been able to keep up with current technologies. This report surveys the major techniques used in industry and academia to ensure reliability in large and critical computer systems.

  5. Technical specifications requirements: Automated reasoning applications

    International Nuclear Information System (INIS)

    Lidsky, L.M.; Dobrzeniecki, A.B.

    1990-03-01

    Several software systems were developed and tested to determine what advantages could be gained from explicitly translating complicated regulatory requirements into computerized relationships. The Technical Specifications for US nuclear power plants were chosen as the test-bed application domain, and two analysis systems were developed to monitor plant compliance with operational limits, and track and schedule equipment test and maintenance activities mandated by Technical Specifications. Choosing PROLOG as the computer language to represent these regulatory requirements resulted in a natural match between the semantic structure of the written specifications and the corollary coded rules. Additional research results affirmed the utility of declarative programming styles, explicit management of problem complexity, and attention to the robustness and flexibility of the overall software systems. 5 refs., 2 figs

  6. Earth Observing System (EOS)/ Advanced Microwave Sounding Unit-A (AMSU-A): Special Test Equipment. Software Requirements

    Science.gov (United States)

    Schwantje, Robert

    1995-01-01

    This document defines the functional, performance, and interface requirements for the Earth Observing System/Advanced Microwave Sounding Unit-A (EOS/AMSU-A) Special Test Equipment (STE) software used in the test and integration of the instruments.

  7. Agile Software Development

    Science.gov (United States)

    Biju, Soly Mathew

    2008-01-01

    Many software development firms are now adopting the agile software development method. This method involves the customer at every level of software development, thus reducing the impact of change in the requirement at a later stage. In this article, the principles of the agile method for software development are explored and there is a focus on…

  8. Study on the BES Ⅲ offline software performance

    International Nuclear Information System (INIS)

    Zhang Xiaomei; Sun Gongxing

    2011-01-01

    Performance monitor and analysis on the BESⅢ offline software system is very useful for the software optimization and the improvement of CPU and memory usage. It presented a feasible performance monitoring service based on GAUDI, and carried out performance tests and analysis on the BESⅢ simulation and reconstruction with the service. (authors)

  9. Software Assurance Competency Model

    Science.gov (United States)

    2013-03-01

    COTS) software , and software as a service ( SaaS ). L2: Define and analyze risks in the acquisition of contracted software , COTS software , and SaaS ...2010a]: Application of technologies and processes to achieve a required level of confidence that software systems and services function in the...

  10. Application of neural network and pattern recognition software to the automated analysis of continuous nuclear monitoring of on-load reactors

    Energy Technology Data Exchange (ETDEWEB)

    Howell, J.A.; Eccleston, G.W.; Halbig, J.K.; Klosterbuer, S.F. [Los Alamos National Lab., NM (United States); Larson, T.W. [California Polytechnic State Univ., San Luis Obispo, CA (US)

    1993-08-01

    Automated analysis using pattern recognition and neural network software can help interpret data, call attention to potential anomalies, and improve safeguards effectiveness. Automated software analysis, based on pattern recognition and neural networks, was applied to data collected from a radiation core discharge monitor system located adjacent to an on-load reactor core. Unattended radiation sensors continuously collect data to monitor on-line refueling operations in the reactor. The huge volume of data collected from a number of radiation channels makes it difficult for a safeguards inspector to review it all, check for consistency among the measurement channels, and find anomalies. Pattern recognition and neural network software can analyze large volumes of data from continuous, unattended measurements, thereby improving and automating the detection of anomalies. The authors developed a prototype pattern recognition program that determines the reactor power level and identifies the times when fuel bundles are pushed through the core during on-line refueling. Neural network models were also developed to predict fuel bundle burnup to calculate the region on the on-load reactor face from which fuel bundles were discharged based on the radiation signals. In the preliminary data set, which was limited and consisted of four distinct burnup regions, the neural network model correctly predicted the burnup region with an accuracy of 92%.

  11. Requirements to a Norwegian national automatic gamma monitoring system

    DEFF Research Database (Denmark)

    Lauritzen, B.; Jensen, Per Hedemann; Nielsen, F.

    2005-01-01

    increments above the natural background levels. The study is based upon simplified deterministic calculations of the radiological consequences of generic nuclear accident scenarios. The density of gammamonitoring stations has been estimated from an analysis of the dispersion of radioactive materials over......An assessment of the overall requirements to a Norwegian gamma-monitoring network is undertaken with special emphasis on the geographical distribution of automatic gamma monitoring stations, type of detectors in such stations and the sensitivity of thesystem in terms of ambient dose equivalent rate...

  12. Approaches that use software to support the prevention of pressure ulcer: A systematic review.

    Science.gov (United States)

    Marchione, F G; Araújo, L M Q; Araújo, L V

    2015-10-01

    The incidence and costs for pressure ulcer (PU) treatment remain high even though preventive methods are applied. Approaches that use software to support the prevention of PU are presented in the literature to make it more effective. Identify the state of art of the approaches that use software to support the prevention of PUs. A systematic literature review was performed to analyze approaches that use software to support the prevention of PU. ACM, IEEE, PubMed, Scopus, CINAHL and Embase databases have been searched with a predetermined search string to identify primary studies. We selected the ones that met the established inclusion criteria. Thirty-six articles met the inclusion criteria. To support prevention, most approaches monitor the patient to provide information about exposure to pressure, temperature level, humidity level and estimated body position in bed providing risk factor intensity charts and intensity maps. The main method to perform patient's monitoring is using sensors installed on the mattress, but recently, alternative methods have been proposed such as electronic sensors and tactile sensory coils. Part of the approaches performs automated management of the risk factors using ventilation tubes and mattresses with porous cells to decrease body's temperature and movable cells to automatically redistribute the pressure over the body. Matters as cost of the approach, patient comfort and hygiene of the monitoring equipment is only briefly discussed in the selected articles. No experiments have been conducted to evidence the approached may reduce PU incidence. Currently, approaches that use software to support the prevention of PU provide relevant information to health professionals such as risk factor intensity charts and intensity maps. Some of them can even automatically manage risk factors in a limited way. Yet, the approaches are based on risk factor monitoring methods that require patient's contact with the monitoring equipment. Therefore, some

  13. Earth Science Informatics Community Requirements for Improving Sustainable Science Software Practices: User Perspectives and Implications for Organizational Action

    Science.gov (United States)

    Downs, R. R.; Lenhardt, W. C.; Robinson, E.

    2014-12-01

    Science software is integral to the scientific process and must be developed and managed in a sustainable manner to ensure future access to scientific data and related resources. Organizations that are part of the scientific enterprise, as well as members of the scientific community who work within these entities, can contribute to the sustainability of science software and to practices that improve scientific community capabilities for science software sustainability. As science becomes increasingly digital and therefore, dependent on software, improving community practices for sustainable science software will contribute to the sustainability of science. Members of the Earth science informatics community, including scientific data producers and distributers, end-user scientists, system and application developers, and data center managers, use science software regularly and face the challenges and the opportunities that science software presents for the sustainability of science. To gain insight on practices needed for the sustainability of science software from the science software experiences of the Earth science informatics community, an interdisciplinary group of 300 community members were asked to engage in simultaneous roundtable discussions and report on their answers to questions about the requirements for improving scientific software sustainability. This paper will present an analysis of the issues reported and the conclusions offered by the participants. These results provide perspectives for science software sustainability practices and have implications for actions that organizations and their leadership can initiate to improve the sustainability of science software.

  14. THE TECHNIQUE OF ANALYSIS OF SOFTWARE OF ON-BOARD COMPUTERS OF AIR VESSEL TO ABSENCE OF UNDECLARED CAPABILITIES BY SIGNATURE-HEURISTIC WAY

    Directory of Open Access Journals (Sweden)

    Viktor Ivanovich Petrov

    2017-01-01

    Full Text Available The article considers the issues of civil aviation aircraft onboard computers data safety. Infor- mation security undeclared capabilities stand for technical equipment or software possibilities, which are not mentioned in the documentation. Documentation and tests content requirements are imposed during the software certification. Documentation requirements include documents composition and content of control (specification, description and program code, the source code. Test requirements include: static analysis of program codes (including the compliance of the sources with their loading modules monitoring; dynamic analysis of source code (including implementation of routes monitor- ing. Currently, there are no complex measures for checking onboard computer software. There are no rules and regulations that can allow controlling foreign production aircraft software, and the actual receiving of software is difficult. Consequently, the author suggests developing the basics of aviation rules and regulations, which allow to analyze the programs of CA aircraft onboard computers. If there are no software source codes the two approaches of code analysis are used: a structural static and dy- namic analysis of the source code; signature-heuristic analysis of potentially dangerous operations. Static analysis determines the behavior of the program by reading the program code (without running the program which is represented in the assembler language - disassembly listing. Program tracing is performed by the dynamic analysis. The analysis of aircraft software ability to detect undeclared capa- bilities using the interactive disassembler was considered in this article.

  15. Real time software for the control and monitoring of DIII-D system interlocks

    International Nuclear Information System (INIS)

    Broesch, J.D.; Penaflor, B.G.; Coon, R.M.; Harris, J.J.; Scoville, J.T.

    1996-10-01

    This paper describes the real time, multi-tasking, multi-user software and communications of the E-Power Supply System Integrated Controller (EPSSIC) for the DIII-D tokamak. EPSSIC performs the DIII-D system wide go/no-go determination for the plasma sequencing. This paper discusses the data module handling, task work load balancing, and communications requirements. Operational experience with the new EPSSIC and recent improvements to this system are also described

  16. System of message for gamma-radiation monitor

    International Nuclear Information System (INIS)

    Bolic, M.D.; Koturovic, A.M.

    2001-01-01

    Paper describes a system of voice messages for gamma-radiation monitor based on PC. The systems reproduces recorded messages that is simpler than the process of their synthesis. Message choice is based on combination of recorded digital results and/or received reference messages or warnings. The system of generation of voice messages applies the Windows based software. The total memory array required to create independent voice system is maximum 1.7 mbyte. The monitor may be used for continuous monitoring of radioactivity level with 5-8 s period of message repetition. Another option of the system operation is based on monitor application for the environment monitoring. Period of messages in this case is equal to 5-30 min [ru

  17. Framework for Small-Scale Experiments in Software Engineering: Guidance and Control Software Project: Software Engineering Case Study

    Science.gov (United States)

    Hayhurst, Kelly J.

    1998-01-01

    Software is becoming increasingly significant in today's critical avionics systems. To achieve safe, reliable software, government regulatory agencies such as the Federal Aviation Administration (FAA) and the Department of Defense mandate the use of certain software development methods. However, little scientific evidence exists to show a correlation between software development methods and product quality. Given this lack of evidence, a series of experiments has been conducted to understand why and how software fails. The Guidance and Control Software (GCS) project is the latest in this series. The GCS project is a case study of the Requirements and Technical Concepts for Aviation RTCA/DO-178B guidelines, Software Considerations in Airborne Systems and Equipment Certification. All civil transport airframe and equipment vendors are expected to comply with these guidelines in building systems to be certified by the FAA for use in commercial aircraft. For the case study, two implementations of a guidance and control application were developed to comply with the DO-178B guidelines for Level A (critical) software. The development included the requirements, design, coding, verification, configuration management, and quality assurance processes. This paper discusses the details of the GCS project and presents the results of the case study.

  18. Hybrid causal methodology and software platform for probabilistic risk assessment and safety monitoring of socio-technical systems

    Energy Technology Data Exchange (ETDEWEB)

    Groth, Katrina, E-mail: kgroth@umd.ed [Center for Risk and Reliability, 0151 Glenn L. Martin Hall, University of Maryland, College Park, MD 20742 (United States); Wang Chengdong; Mosleh, Ali [Center for Risk and Reliability, 0151 Glenn L. Martin Hall, University of Maryland, College Park, MD 20742 (United States)

    2010-12-15

    This paper introduces an integrated framework and software platform for probabilistic risk assessment (PRA) and safety monitoring of complex socio-technical systems. An overview of the three-layer hybrid causal logic (HCL) modeling approach and corresponding algorithms, implemented in the Trilith software platform, are provided. The HCL approach enhances typical PRA methods by quantitatively including the influence of soft causal factors introduced by human and organizational aspects of a system. The framework allows different modeling techniques to be used for different aspects of the socio-technical system. The HCL approach combines the power of traditional event sequence diagram (ESD)event tree (ET) and fault tree (FT) techniques for modeling deterministic causal paths, with the flexibility of Bayesian belief networks for modeling non-deterministic cause-effect relationships among system elements (suitable for modeling human and organizational influences). Trilith enables analysts to construct HCL models and perform quantitative risk assessment and management of complex systems. The risk management capabilities included are HCL-based risk importance measures, hazard identification and ranking, precursor analysis, safety indicator monitoring, and root cause analysis. This paper describes the capabilities of the Trilith platform and power of the HCL algorithm by use of example risk models for a type of aviation accident (aircraft taking off from the wrong runway).

  19. Hybrid causal methodology and software platform for probabilistic risk assessment and safety monitoring of socio-technical systems

    International Nuclear Information System (INIS)

    Groth, Katrina; Wang Chengdong; Mosleh, Ali

    2010-01-01

    This paper introduces an integrated framework and software platform for probabilistic risk assessment (PRA) and safety monitoring of complex socio-technical systems. An overview of the three-layer hybrid causal logic (HCL) modeling approach and corresponding algorithms, implemented in the Trilith software platform, are provided. The HCL approach enhances typical PRA methods by quantitatively including the influence of soft causal factors introduced by human and organizational aspects of a system. The framework allows different modeling techniques to be used for different aspects of the socio-technical system. The HCL approach combines the power of traditional event sequence diagram (ESD)event tree (ET) and fault tree (FT) techniques for modeling deterministic causal paths, with the flexibility of Bayesian belief networks for modeling non-deterministic cause-effect relationships among system elements (suitable for modeling human and organizational influences). Trilith enables analysts to construct HCL models and perform quantitative risk assessment and management of complex systems. The risk management capabilities included are HCL-based risk importance measures, hazard identification and ranking, precursor analysis, safety indicator monitoring, and root cause analysis. This paper describes the capabilities of the Trilith platform and power of the HCL algorithm by use of example risk models for a type of aviation accident (aircraft taking off from the wrong runway).

  20. Software Prototyping: A Case Report of Refining User Requirements for a Health Information Exchange Dashboard.

    Science.gov (United States)

    Nelson, Scott D; Del Fiol, Guilherme; Hanseler, Haley; Crouch, Barbara Insley; Cummins, Mollie R

    2016-01-01

    Health information exchange (HIE) between Poison Control Centers (PCCs) and Emergency Departments (EDs) could improve care of poisoned patients. However, PCC information systems are not designed to facilitate HIE with EDs; therefore, we are developing specialized software to support HIE within the normal workflow of the PCC using user-centered design and rapid prototyping. To describe the design of an HIE dashboard and the refinement of user requirements through rapid prototyping. Using previously elicited user requirements, we designed low-fidelity sketches of designs on paper with iterative refinement. Next, we designed an interactive high-fidelity prototype and conducted scenario-based usability tests with end users. Users were asked to think aloud while accomplishing tasks related to a case vignette. After testing, the users provided feedback and evaluated the prototype using the System Usability Scale (SUS). Survey results from three users provided useful feedback that was then incorporated into the design. After achieving a stable design, we used the prototype itself as the specification for development of the actual software. Benefits of prototyping included having 1) subject-matter experts heavily involved with the design; 2) flexibility to make rapid changes, 3) the ability to minimize software development efforts early in the design stage; 4) rapid finalization of requirements; 5) early visualization of designs; 6) and a powerful vehicle for communication of the design to the programmers. Challenges included 1) time and effort to develop the prototypes and case scenarios; 2) no simulation of system performance; 3) not having all proposed functionality available in the final product; and 4) missing needed data elements in the PCC information system.

  1. Software V and V of PPS for Shin-Hanul Nuclear Power Plant Units 1 and 2

    International Nuclear Information System (INIS)

    Park, Cheollak; Kang, Dongpa; Choe, Changhui; Sohn, Sedo; Beak, Seungmin

    2013-01-01

    Software V and V processes determine whether the development products of a given activity conform to the requirements of that activity and whether the software satisfies its intended use and user needs. This paper introduces the software V and V activities and tasks performed during the software development life cycle performed during the software development life cycle of the Plant Protection System (PPS) for Shin-Hanul Nuclear Power Plant Units 1 and 2 (SHN 1 and 2). The PPS generates signals to actuate Reactor Trip (RT) and Engineered Safety Features (ESF) whenever monitored processes exceed predetermined limits, and the PPS software is classified safety critical and an independent V and V is thus required according to regulations, code and standards. The software V and V efforts, sufficiently disciplined and rigorous, are quite essential to demonstrate that the software development process is of a high quality. The software V and V of PPS for SHN 1 and 2 has been accomplished successfully with systematic V and V procedures and methods established until test phase in compliance with related code and standards. In particular, the use of automated tools such as LDRA and DOORS greatly has contributed to an improvement of a software quality, and a reduction of a verification time and human errors

  2. Software Architecture Evolution

    Science.gov (United States)

    Barnes, Jeffrey M.

    2013-01-01

    Many software systems eventually undergo changes to their basic architectural structure. Such changes may be prompted by new feature requests, new quality attribute requirements, changing technology, or other reasons. Whatever the causes, architecture evolution is commonplace in real-world software projects. Today's software architects, however,…

  3. Software factory techniques applied to Process Control at CERN

    CERN Multimedia

    Dutour, MD

    2007-01-01

    The CERN Large Hadron Collider (LHC) requires constant monitoring and control of quantities of parameters to guarantee operational conditions. For this purpose, a methodology called UNICOS (UNIfied Industrial COntrols Systems) has been implemented to standardize the design of process control applications. To further accelerate the development of these applications, we migrated our existing UNICOS tooling suite toward a software factory in charge of assembling project, domain and technical information seamlessly into deployable PLC (Programmable logic Controller) – SCADA (Supervisory Control And Data Acquisition) systems. This software factory delivers consistently high quality by reducing human error and repetitive tasks, and adapts to user specifications in a cost-efficient way. Hence, this production tool is designed to encapsulate and hide the PLC and SCADA target platforms, enabling the experts to focus on the business model rather than specific syntaxes and grammars. Based on industry standard software...

  4. A study of software safety analysis system for safety-critical software

    International Nuclear Information System (INIS)

    Chang, H. S.; Shin, H. K.; Chang, Y. W.; Jung, J. C.; Kim, J. H.; Han, H. H.; Son, H. S.

    2004-01-01

    The core factors and requirements for the safety-critical software traced and the methodology adopted in each stage of software life cycle are presented. In concept phase, Failure Modes and Effects Analysis (FMEA) for the system has been performed. The feasibility evaluation of selected safety parameter was performed and Preliminary Hazards Analysis list was prepared using HAZOP(Hazard and Operability) technique. And the check list for management control has been produced via walk-through technique. Based on the evaluation of the check list, activities to be performed in requirement phase have been determined. In the design phase, hazard analysis has been performed to check the safety capability of the system with regard to safety software algorithm using Fault Tree Analysis (FTA). In the test phase, the test items based on FMEA have been checked for fitness guided by an accident scenario. The pressurizer low pressure trip algorithm has been selected to apply FTA method to software safety analysis as a sample. By applying CASE tool, the requirements traceability of safety critical system has been enhanced during all of software life cycle phases

  5. Functions and requirements for Hanford single-shell tank leakage detection and monitoring

    International Nuclear Information System (INIS)

    Cruse, J.M.; Ohl, P.C.

    1995-01-01

    This document provides the initial functions and requirements for leakage detection and monitoring applicable to past and potential future leakage from the Hanford Site's 149 single-shell high-level waste tanks. This mission is a part of the overall mission of the Westinghouse Hanford Company Tank Waste Remediation System division to remediate the tank waste in a safe and acceptable manner. Systems engineering principles are being applied to this effort. This document reflects the an initial step in the systems engineering approach to decompose the mission into primary functions and requirements. The document is considered approximately 30% complete relative to the effort required to produce a final version that can be used to support demonstration and/or procurement of technologies. The functions and requirements in this document apply to detection and monitoring of below ground leaks from SST containment boundaries and the resulting soil contamination. Leakage detection and monitoring is invoked in the TWRS Program in three fourth level functions: (1) Store Waste, (2) Retrieve Waste, and (3) Disposition Excess Facilities (as identified in DOE/RL-92-60 Rev. 1, Tank Waste Remediation System Functions and Requirements)

  6. 40 CFR 98.164 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... Methods for Instrumental Determination of Carbon, Hydrogen, and Nitrogen in Petroleum Products and... Determination of Carbon, Hydrogen, and Nitrogen in Laboratory Samples of Coal (incorporated by reference, see... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Hydrogen Production § 98.164 Monitoring and QA/QC requirements...

  7. Research of hydroelectric generating set low-frequency vibration monitoring system based on optical fiber sensing

    Science.gov (United States)

    Min, Li; Zhang, Xiaolei; Zhang, Faxiang; Sun, Zhihui; Li, ShuJuan; Wang, Meng; Wang, Chang

    2017-10-01

    In order to satisfy hydroelectric generating set low-frequency vibration monitoring, the design of Passive low-frequency vibration monitoring system based on Optical fiber sensing in this paper. The hardware of the system adopts the passive optical fiber grating sensor and unbalanced-Michelson interferometer. The software system is used to programming by Labview software and finishing the control of system. The experiment show that this system has good performance on the standard vibration testing-platform and it meets system requirements. The frequency of the monitoring system can be as low as 0.2Hz and the resolution is 0.01Hz.

  8. Clinical evaluation of a dose monitoring software tool based on Monte Carlo Simulation in assessment of eye lens doses for cranial CT scans

    Energy Technology Data Exchange (ETDEWEB)

    Guberina, Nika; Suntharalingam, Saravanabavaan; Nassenstein, Kai; Forsting, Michael; Theysohn, Jens; Wetter, Axel; Ringelstein, Adrian [University Hospital Essen, Institute of Diagnostic and Interventional Radiology and Neuroradiology, Essen (Germany)

    2016-10-15

    The aim of this study was to verify the results of a dose monitoring software tool based on Monte Carlo Simulation (MCS) in assessment of eye lens doses for cranial CT scans. In cooperation with the Federal Office for Radiation Protection (Neuherberg, Germany), phantom measurements were performed with thermoluminescence dosimeters (TLD LiF:Mg,Ti) using cranial CT protocols: (I) CT angiography; (II) unenhanced, cranial CT scans with gantry angulation at a single and (III) without gantry angulation at a dual source CT scanner. Eye lens doses calculated by the dose monitoring tool based on MCS and assessed with TLDs were compared. Eye lens doses are summarized as follows: (I) CT angiography (a) MCS 7 mSv, (b) TLD 5 mSv; (II) unenhanced, cranial CT scan with gantry angulation, (c) MCS 45 mSv, (d) TLD 5 mSv; (III) unenhanced, cranial CT scan without gantry angulation (e) MCS 38 mSv, (f) TLD 35 mSv. Intermodality comparison shows an inaccurate calculation of eye lens doses in unenhanced cranial CT protocols at the single source CT scanner due to the disregard of gantry angulation. On the contrary, the dose monitoring tool showed an accurate calculation of eye lens doses at the dual source CT scanner without gantry angulation and for CT angiography examinations. The dose monitoring software tool based on MCS gave accurate estimates of eye lens doses in cranial CT protocols. However, knowledge of protocol and software specific influences is crucial for correct assessment of eye lens doses in routine clinical use. (orig.)

  9. Clinical evaluation of a dose monitoring software tool based on Monte Carlo Simulation in assessment of eye lens doses for cranial CT scans

    International Nuclear Information System (INIS)

    Guberina, Nika; Suntharalingam, Saravanabavaan; Nassenstein, Kai; Forsting, Michael; Theysohn, Jens; Wetter, Axel; Ringelstein, Adrian

    2016-01-01

    The aim of this study was to verify the results of a dose monitoring software tool based on Monte Carlo Simulation (MCS) in assessment of eye lens doses for cranial CT scans. In cooperation with the Federal Office for Radiation Protection (Neuherberg, Germany), phantom measurements were performed with thermoluminescence dosimeters (TLD LiF:Mg,Ti) using cranial CT protocols: (I) CT angiography; (II) unenhanced, cranial CT scans with gantry angulation at a single and (III) without gantry angulation at a dual source CT scanner. Eye lens doses calculated by the dose monitoring tool based on MCS and assessed with TLDs were compared. Eye lens doses are summarized as follows: (I) CT angiography (a) MCS 7 mSv, (b) TLD 5 mSv; (II) unenhanced, cranial CT scan with gantry angulation, (c) MCS 45 mSv, (d) TLD 5 mSv; (III) unenhanced, cranial CT scan without gantry angulation (e) MCS 38 mSv, (f) TLD 35 mSv. Intermodality comparison shows an inaccurate calculation of eye lens doses in unenhanced cranial CT protocols at the single source CT scanner due to the disregard of gantry angulation. On the contrary, the dose monitoring tool showed an accurate calculation of eye lens doses at the dual source CT scanner without gantry angulation and for CT angiography examinations. The dose monitoring software tool based on MCS gave accurate estimates of eye lens doses in cranial CT protocols. However, knowledge of protocol and software specific influences is crucial for correct assessment of eye lens doses in routine clinical use. (orig.)

  10. An application study for the class 1E digital control and monitoring system

    International Nuclear Information System (INIS)

    Hiroyuki Fukumitsu

    1998-01-01

    This paper presents an application study for the Class 1E digital control and monitoring system to the next Japanese plants, especially about MMIS. The system architecture of hardware and software is also introduced, which will explain the strategic plan for the necessary software verification and validation according to the latest requirement from Japanese regulatory guide. (author)

  11. 48 CFR 227.7203-2 - Acquisition of noncommercial computer software and computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... at one site or multiple site licenses, and the format and media in which the software or... noncommercial computer software and computer software documentation. 227.7203-2 Section 227.7203-2 Federal... CONTRACTING REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software...

  12. A software architecture for a transportation control tower

    NARCIS (Netherlands)

    Baumgrass, A.; Dijkman, R.M.; Grefen, P.W.P.J.; Pourmirza, S.; Völzer, H.; Weske, M.H.

    2014-01-01

    A Transportation Control Tower is a software application that facilitates transportation planners with easily monitoring and dispatching transportation resources. This paper presents a software architecture for such an application. It focuses in particular on the novel aspects of the software

  13. Regulation E 69-14. Monitoring requirements for medical devices

    International Nuclear Information System (INIS)

    2015-01-01

    In the 'Regulations for the State Evaluation and Registration of Medical Equipment' force (Hereinafter Rules) set forth in Chapter VII, Articles 79 and 86, the monitoring activity as one of the programs necessary for evaluating the safety and effectiveness of medical monitoring equipment. In the years 2008 and 2011 were approved and implemented by the Center for State Control of Medical Equipment (CCEEM) Regulations and -1.1 ER and ER-1 that support and regulatory requirements 'Control and monitoring of pacemakers and implantable defibrillators' and 'Assessment, recording and control after market surgical silicone implants,' which are specific to these products and have provided a useful result for the performance of the activity. Given the number and diversity of high-risk medical devices as implantable or sustain human life that are brought into our National Health System (SNS), a regulation establishing control over the behavior becomes necessary safety and effectiveness of this equipment during use, which provide inputs to risk management. The objective of this regulation is to establish the regulatory requirements for tracking medical equipment introduced in the NHS. The provisions of this Regulation is aimed at health institutions, to CECMED as manufacturers, suppliers, distributors and importers of medical equipment.

  14. Operational intervention levels and related requirements on radiation monitoring during pre-release / release phase of an accident

    International Nuclear Information System (INIS)

    Carny, P.; Cabanekova, H

    2003-01-01

    In this paper authors discusses required outputs of emergency radiological monitoring in various phases of an accident and rationale of these requirements. In various phases of an accident various intervention levels are important and consequently various radiological quantities should be preferably measured. Distinguished tasks or aims of monitoring in different phases of accident have strong influence on methods of monitoring, instrumentation and capabilities of monitoring groups. Required tasks and outputs of monitoring are discussed

  15. ATLAS Future Framework Requirements Group Report

    CERN Document Server

    The ATLAS collaboration

    2016-01-01

    The Future Frameworks Requirements Group was constituted in Summer 2013 to consider and summarise the framework requirements from trigger and offline for configuring, scheduling and monitoring the data processing software needed by the ATLAS experiment. The principal motivation for such a re-examination arises from the current and anticipated evolution of CPUs, where multiple cores, hyper-threading and wide vector registers require a shift to a concurrent programming model. Such a model requires extensive changes in the current Gaudi/Athena frameworks and offers the opportunity to consider how HLT and offline processing can be better accommodated within the ATLAS framework. This note contains the report of the Future Frameworks Requirements Group.

  16. 40 CFR 98.214 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98.214 Section 98.214 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS... standard method or other enhanced industry consensus standard method published by an industry consensus...

  17. Software defined network inference with evolutionary optimal observation matrices

    OpenAIRE

    Malboubi, M; Gong, Y; Yang, Z; Wang, X; Chuah, CN; Sharma, P

    2017-01-01

    © 2017 Elsevier B.V. A key requirement for network management is the accurate and reliable monitoring of relevant network characteristics. In today's large-scale networks, this is a challenging task due to the scarcity of network measurement resources and the hard constraints that this imposes. This paper proposes a new framework, called SNIPER, which leverages the flexibility provided by Software-Defined Networking (SDN) to design the optimal observation or measurement matrix that can lead t...

  18. 40 CFR 63.6125 - What are my monitor installation, operation, and maintenance requirements?

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 12 2010-07-01 2010-07-01 true What are my monitor installation... Combustion Turbines Testing and Initial Compliance Requirements § 63.6125 What are my monitor installation... limitation and you are not using an oxidation catalyst, you must continuously monitor any parameters...

  19. Monitoring of noninvasive ventilation by built-in software of home bilevel ventilators: a bench study.

    Science.gov (United States)

    Contal, Olivier; Vignaux, Laurence; Combescure, Christophe; Pepin, Jean-Louis; Jolliet, Philippe; Janssens, Jean-Paul

    2012-02-01

    Current bilevel positive-pressure ventilators for home noninvasive ventilation (NIV) provide physicians with software that records items important for patient monitoring, such as compliance, tidal volume (Vt), and leaks. However, to our knowledge, the validity of this information has not yet been independently assessed. Testing was done for seven home ventilators on a bench model adapted to simulate NIV and generate unintentional leaks (ie, other than of the mask exhalation valve). Five levels of leaks were simulated using a computer-driven solenoid valve (0-60 L/min) at different levels of inspiratory pressure (15 and 25 cm H(2)O) and at a fixed expiratory pressure (5 cm H(2)O), for a total of 10 conditions. Bench data were compared with results retrieved from ventilator software for leaks and Vt. For assessing leaks, three of the devices tested were highly reliable, with a small bias (0.3-0.9 L/min), narrow limits of agreement (LA), and high correlations (R(2), 0.993-0.997) when comparing ventilator software and bench results; conversely, for four ventilators, bias ranged from -6.0 L/min to -25.9 L/min, exceeding -10 L/min for two devices, with wide LA and lower correlations (R(2), 0.70-0.98). Bias for leaks increased markedly with the importance of leaks in three devices. Vt was underestimated by all devices, and bias (range, 66-236 mL) increased with higher insufflation pressures. Only two devices had a bias ventilation must be aware of differences in the estimation of leaks and Vt by ventilator software. Also, leaks are reported in different ways according to the device used.

  20. Software Authentication

    International Nuclear Information System (INIS)

    Wolford, J.K.; Geelhood, B.D.; Hamilton, V.A.; Ingraham, J.; MacArthur, D.W.; Mitchell, D.J.; Mullens, J.A.; Vanier, P. E.; White, G.K.; Whiteson, R.

    2001-01-01

    The effort to define guidance for authentication of software for arms control and nuclear material transparency measurements draws on a variety of disciplines and has involved synthesizing established criteria and practices with newer methods. Challenges include the need to protect classified information that the software manipulates as well as deal with the rapid pace of innovation in the technology of nuclear material monitoring. The resulting guidance will shape the design of future systems and inform the process of authentication of instruments now being developed. This paper explores the technical issues underlying the guidance and presents its major tenets

  1. Monitoring and Detection Platform to Prevent Anomalous Situations in Home Care

    Directory of Open Access Journals (Sweden)

    Gabriel Villarrubia

    2014-06-01

    Full Text Available Monitoring and tracking people at home usually requires high cost hardware installations, which implies they are not affordable in many situations. This study/paper proposes a monitoring and tracking system for people with medical problems. A virtual organization of agents based on the PANGEA platform, which allows the easy integration of different devices, was created for this study. In this case, a virtual organization was implemented to track and monitor patients carrying a Holter monitor. The system includes the hardware and software required to perform: ECG measurements, monitoring through accelerometers and WiFi networks. Furthermore, the use of interactive television can moderate interactivity with the user. The system makes it possible to merge the information and facilitates patient tracking efficiently with low cost.

  2. Wireless data management system for environmental monitoring in livestock buildings

    Directory of Open Access Journals (Sweden)

    James Gray

    2017-03-01

    Full Text Available The impact of air quality on the health, welfare and productivity of livestock needs to be considered, especially when livestock are kept in enclosed buildings. The monitoring of such environmental factors allows for the development of appropriate strategies to reduce detrimental effects of sub-optimal air quality on the respiratory health of both livestock and farmers. In 2009, an environmental monitoring system was designed, developed and tested that allowed for the monitoring of a number of airborne pollutants. One limitation of the system was the manual collection of logged data from each unit. This paper identifies limitations of the current environmental monitoring system and suggests a range of networking technologies that can be used to increase usability. Consideration is taken for the networking of environmental monitoring units, as well as the collection of recorded data. Furthermore, the design and development of a software system that is used to collate and store recorded environmental data from multiple farms is explored. In order to design such a system, simplified software engineering processes and methodologies have been utilised. The main steps taken in order to complete the project were requirements elicitation with clients, requirements analysis, system design, implementation and finally testing. The outcome of the project provided a potential prototype for improving the environmental monitoring system and analysis informing the benefit of the implementation.

  3. Practical support for Lean Six Sigma software process definition using IEEE software engineering standards

    CERN Document Server

    Land, Susan K; Walz, John W

    2012-01-01

    Practical Support for Lean Six Sigma Software Process Definition: Using IEEE Software Engineering Standards addresses the task of meeting the specific documentation requirements in support of Lean Six Sigma. This book provides a set of templates supporting the documentation required for basic software project control and management and covers the integration of these templates for their entire product development life cycle. Find detailed documentation guidance in the form of organizational policy descriptions, integrated set of deployable document templates, artifacts required in suppo

  4. 40 CFR 98.394 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98... by a consensus-based standards organization exists, such a method shall be used. Consensus-based... (NAESB). (ii) Where no appropriate standard method developed by a consensus-based standards organization...

  5. 40 CFR 98.424 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98... published by a consensus-based standards organization if such a method exists. Consensus-based standards...). (ii) Where no appropriate standard method developed by a consensus-based standards organization exists...

  6. Software as a service approach to sensor simulation software deployment

    Science.gov (United States)

    Webster, Steven; Miller, Gordon; Mayott, Gregory

    2012-05-01

    Traditionally, military simulation has been problem domain specific. Executing an exercise currently requires multiple simulation software providers to specialize, deploy, and configure their respective implementations, integrate the collection of software to achieve a specific system behavior, and then execute for the purpose at hand. This approach leads to rigid system integrations which require simulation expertise for each deployment due to changes in location, hardware, and software. Our alternative is Software as a Service (SaaS) predicated on the virtualization of Night Vision Electronic Sensors (NVESD) sensor simulations as an exemplary case. Management middleware elements layer self provisioning, configuration, and integration services onto the virtualized sensors to present a system of services at run time. Given an Infrastructure as a Service (IaaS) environment, enabled and managed system of simulations yields a durable SaaS delivery without requiring user simulation expertise. Persistent SaaS simulations would provide on demand availability to connected users, decrease integration costs and timelines, and benefit the domain community from immediate deployment of lessons learned.

  7. Software life after in-service

    International Nuclear Information System (INIS)

    Tseng, M.; Eng, P.

    1993-01-01

    Software engineers and designers tend to conclude a software project at the in-service milestone of the software life cycle. But the reality is that the 'life after in-service' is significantly longer than other phases of the life cycle, typically 20 years or more depending on the maintainability of the hardware platform and the designed life of the plant. During this period, the software asset (as with other physical assets in the plant) continues to be upgraded to correct deficiencies, meet new requirements, cope with obsolescence of equipment and so on. The software life cycle ends with a migration of the software to a different platform. It is typical in a software development project to put a great deal of emphasis on design methodologies, techniques, tools, development environment, standard procedures, and project management to ensure quality product is delivered on schedule and within budget. More often than not, a disproportion of emphasis is placed on the issues and needs of the in-service phase. Once the software is in-service, the designers move on to other projects, while the maintenance and support staff must manage the software. This paper examines the issues in three steps. First it presents a view of software from maintenance and support staff perspectives, including complexity of software, suitability of documentation, configuration management, training, difficulties and risks associated with making changes, required skills and knowledge. Second, it identifies the concerns raised from these viewpoints, including costs of maintaining the software, ability to meet additional requirements, availability of support tools, length of time required to engineer and install changes, and a strategy for the migration of software asset. Finally it discusses some approaches to deal with the concerns. (Author) 5 refs., fig

  8. Simulation-based Testing of Control Software

    Energy Technology Data Exchange (ETDEWEB)

    Ozmen, Ozgur [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Nutaro, James J. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Sanyal, Jibonananda [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Olama, Mohammed M. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2017-02-10

    It is impossible to adequately test complex software by examining its operation in a physical prototype of the system monitored. Adequate test coverage can require millions of test cases, and the cost of equipment prototypes combined with the real-time constraints of testing with them makes it infeasible to sample more than a small number of these tests. Model based testing seeks to avoid this problem by allowing for large numbers of relatively inexpensive virtual prototypes that operate in simulation time at a speed limited only by the available computing resources. In this report, we describe how a computer system emulator can be used as part of a model based testing environment; specifically, we show that a complete software stack including operating system and application software - can be deployed within a simulated environment, and that these simulations can proceed as fast as possible. To illustrate this approach to model based testing, we describe how it is being used to test several building control systems that act to coordinate air conditioning loads for the purpose of reducing peak demand. These tests involve the use of ADEVS (A Discrete Event System Simulator) and QEMU (Quick Emulator) to host the operational software within the simulation, and a building model developed with the MODELICA programming language using Buildings Library and packaged as an FMU (Functional Mock-up Unit) that serves as the virtual test environment.

  9. Monitoring requirements for assessment of internal dose

    International Nuclear Information System (INIS)

    Eckerman, K.F.

    1985-01-01

    Data obtained by routine personnel monitoring is usually not a sufficient basis for estimation of dose. Collected data must be interpreted carefully and supplemented with appropriate information before reasonably accurate estimates of dose (i.e., accurate enough to indicate whether or nor personnel are exposed in excess of recommended limits) can be developed. When the exposure is of sufficient magnitude that a rather precise estimate of dose is needed, the health physicist will bring to bear on the problem other, more refined, methods of dosimetry. These might include a reconstruction of the incident and, for internal emitters, an extensive series of in vivo measurements or analyses of excreta. Thus, cases of special significance must often be evaluated using techniques and resources beyond those routinely employed. This is not a criticism of most routine monitoring programs. These programs are usually carefully designed in a manner commensurate with the degree of exposure routinely encountered and the requirement of a practical program of radiation protection. 10 refs

  10. 40 CFR 98.144 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98... fraction for each carbonate consumed based on sampling and chemical analysis using an industry consensus... testing method published by an industry consensus standards organization (e.g., ASTM, ASME, API, etc.). ...

  11. 40 CFR 98.404 - Monitoring and QA/QC requirements.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98... published by a consensus-based standards organization exists, such a method shall be used. Consensus-based... (NAESB). (ii) Where no appropriate standard method developed by a consensus-based standards organization...

  12. Review of FEWS NET Biophysical Monitoring Requirements

    Science.gov (United States)

    Ross, K. W.; Brown, Molly E.; Verdin, J.; Underwood, L. W.

    2009-01-01

    The Famine Early Warning System Network (FEWS NET) provides monitoring and early warning support to decision makers responsible for responding to famine and food insecurity. FEWS NET transforms satellite remote sensing data into rainfall and vegetation information that can be used by these decision makers. The National Aeronautics and Space Administration has recently funded activities to enhance remote sensing inputs to FEWS NET. To elicit Earth observation requirements, a professional review questionnaire was disseminated to FEWS NET expert end-users: it focused upon operational requirements to determine additional useful remote sensing data and; subsequently, beneficial FEWS NET biophysical supplementary inputs. The review was completed by over 40 experts from around the world, enabling a robust set of professional perspectives to be gathered and analyzed rapidly. Reviewers were asked to evaluate the relative importance of environmental variables and spatio-temporal requirements for Earth science data products, in particular for rainfall and vegetation products. The results showed that spatio-temporal resolution requirements are complex and need to vary according to place, time, and hazard: that high resolution remote sensing products continue to be in demand, and that rainfall and vegetation products were valued as data that provide actionable food security information.

  13. Investigation of Classification and Design Requirements for Digital Software for Advanced Research Reactors

    Energy Technology Data Exchange (ETDEWEB)

    Park, Gee Young; Jung, H. S.; Ryu, J. S.; Park, C

    2005-06-15

    software for use in I and C systems in nuclear power plants and describes the requirements for software development recommended by international standard.

  14. Employing peer-to-peer software distribution in ALICE Grid Services to enable opportunistic use of OSG resources

    CERN Multimedia

    CERN. Geneva; Sakrejda, Iwona

    2012-01-01

    The ALICE Grid infrastructure is based on AliEn, a lightweight open source framework built on Web Services and a Distributed Agent Model in which job agents are submitted onto a grid site to prepare the environment and pull work from a central task queue located at CERN. In the standard configuration, each ALICE grid site supports an ALICE-specific VO box as a single point of contact between the site and the ALICE central services. VO box processes monitor site utilization and job requests (ClusterMonitor), monitor dynamic job and site properties (MonaLisa), perform job agent submission (CE) and deploy job-specific software (PackMan). In particular, requiring a VO box at each site simplifies deployment of job software, done onto a shared file system at the site, and adds redundancy to the overall Grid system. ALICE offline computing, however, has also implemented a peer-to-peer method (based on BitTorrent) for downloading job software directly onto each worker node as needed. By utilizing both this peer-...

  15. LDUA software custodian's notebook

    International Nuclear Information System (INIS)

    Aftanas, B.L.

    1998-01-01

    This plan describes the activities to be performed and controls to be applied to the process of specifying, obtaining, and qualifying the control and data acquisition software for the Light Duty Utility Arm (LDUA) System. It serves the purpose of a software quality assurance plan, a verification and validation plan, and a configuration management plan. This plan applies to all software that is an integral part of the LDUA control and data acquisition system, that is, software that is installed in the computers that are part of the LDUA system as it is deployed in the field. This plan applies to the entire development process, including: requirements; design; implementation; and operations and maintenance. This plan does not apply to any software that is not integral with the LDUA system. This plan has-been prepared in accordance with WHC-CM-6-1 Engineering Practices, EP-2.1; WHC-CM-3-10 Software Practices; and WHC-CM-4-2, QR 19.0, Software Quality Assurance Requirements

  16. Methods of Software Verification

    Directory of Open Access Journals (Sweden)

    R. E. Gurin

    2015-01-01

    Full Text Available This article is devoted to the problem of software verification (SW. Methods of software verification designed to check the software for compliance with the stated requirements such as correctness, system security and system adaptability to small changes in the environment, portability and compatibility, etc. These are various methods both by the operation process and by the way of achieving result. The article describes the static and dynamic methods of software verification and paid attention to the method of symbolic execution. In its review of static analysis are discussed and described the deductive method, and methods for testing the model. A relevant issue of the pros and cons of a particular method is emphasized. The article considers classification of test techniques for each method. In this paper we present and analyze the characteristics and mechanisms of the static analysis of dependencies, as well as their views, which can reduce the number of false positives in situations where the current state of the program combines two or more states obtained both in different paths of execution and in working with multiple object values. Dependences connect various types of software objects: single variables, the elements of composite variables (structure fields, array elements, the size of the heap areas, the length of lines, the number of initialized array elements in the verification code using static methods. The article pays attention to the identification of dependencies within the framework of the abstract interpretation, as well as gives an overview and analysis of the inference tools.Methods of dynamic analysis such as testing, monitoring and profiling are presented and analyzed. Also some kinds of tools are considered which can be applied to the software when using the methods of dynamic analysis. Based on the work a conclusion is drawn, which describes the most relevant problems of analysis techniques, methods of their solutions and

  17. Computing and software

    Directory of Open Access Journals (Sweden)

    White, G. C.

    2004-06-01

    Full Text Available The reality is that the statistical methods used for analysis of data depend upon the availability of software. Analysis of marked animal data is no different than the rest of the statistical field. The methods used for analysis are those that are available in reliable software packages. Thus, the critical importance of having reliable, up–to–date software available to biologists is obvious. Statisticians have continued to develop more robust models, ever expanding the suite of potential analysis methods available. But without software to implement these newer methods, they will languish in the abstract, and not be applied to the problems deserving them. In the Computers and Software Session, two new software packages are described, a comparison of implementation of methods for the estimation of nest survival is provided, and a more speculative paper about how the next generation of software might be structured is presented. Rotella et al. (2004 compare nest survival estimation with different software packages: SAS logistic regression, SAS non–linear mixed models, and Program MARK. Nests are assumed to be visited at various, possibly infrequent, intervals. All of the approaches described compute nest survival with the same likelihood, and require that the age of the nest is known to account for nests that eventually hatch. However, each approach offers advantages and disadvantages, explored by Rotella et al. (2004. Efford et al. (2004 present a new software package called DENSITY. The package computes population abundance and density from trapping arrays and other detection methods with a new and unique approach. DENSITY represents the first major addition to the analysis of trapping arrays in 20 years. Barker & White (2004 discuss how existing software such as Program MARK require that each new model’s likelihood must be programmed specifically for that model. They wishfully think that future software might allow the user to combine

  18. Software testing for evolutionary iterative rapid prototyping

    OpenAIRE

    Davis, Edward V., Jr.

    1990-01-01

    Approved for public release; distribution unlimited. Rapid prototyping is emerging as a promising software development paradigm. It provides a systematic and automatable means of developing a software system under circumstances where initial requirements are not well known or where requirements change frequently during development. To provide high software quality assurance requires sufficient software testing. The unique nature of evolutionary iterative prototyping is not well-suited for ...

  19. Baseline requirements for assessment of mining impact using biological monitoring

    International Nuclear Information System (INIS)

    Humphrey, C.L.; Dostine, P.L.

    1995-01-01

    Biological monitoring programmes for environmental protection should provide for both early detection of possible adverse effects, and assessment of the ecological significance of these effects. Monitoring techniques are required that include responses sensitive to the impact, that can be subjected to rigorous statistical analysis and for which statistical power is high. Such issues in baseline research of 'what and how to measure?' and 'for how long?' have been the focus of a programme being developed to monitor and assess effects of mining operations on the essentially pristine, freshwater ecosystems of the Alligator Rivers Region (ARR) in tropical northern Australia. Application of the BACIP (Before, After, Control, Impact, Paired differences) design, utilizing a form of temporal replication, to univariate (single species) and multivariate (community) data is described. The BACIP design incorporates data from single control and impact sites. We argue for modification of the design for particular studies conducted in streams, to incorporate additional independent control sites from adjacent catchment. Inferential power, by way of (i) more confidently attributing cause to an observed change and (ii) providing information about the ecological significance of the change, will be enhanced using a modified BACIP design. In highly valued environments such as the ARR, monitoring programmes require application of statistical tests with high power to guarantee that an impact no greater than a prescribed amount has gone undetected. A minimum number of baseline years using the BACIP approach would therefore be required in order to achieve some desired level of statistical power. This paper describes the results of power analyses conducted on 2-5 years (depending upon the technique) of baseline data from streams of the ARR and discuss the implications of these results for management. 44 refs., 1 tab., 3 figs

  20. Monitoring and control requirement definition study for Dispersed Storage and Generation (DSG). Volume 2, appendix A: Selected DSG technologies and their general control requirements

    Science.gov (United States)

    1980-01-01

    A consistent approach was sought for both hardware and software which will handle the monitoring and control necessary to integrate a number of different DSG technologies into a common distribution dispatch network. It appears that the control of each of the DSG technologies is compatible with a supervisory control method of operation that lends itself to remote control from a distribution dispatch center.

  1. Copilot: Monitoring Embedded Systems

    Science.gov (United States)

    Pike, Lee; Wegmann, Nis; Niller, Sebastian; Goodloe, Alwyn

    2012-01-01

    Runtime verification (RV) is a natural fit for ultra-critical systems, where correctness is imperative. In ultra-critical systems, even if the software is fault-free, because of the inherent unreliability of commodity hardware and the adversity of operational environments, processing units (and their hosted software) are replicated, and fault-tolerant algorithms are used to compare the outputs. We investigate both software monitoring in distributed fault-tolerant systems, as well as implementing fault-tolerance mechanisms using RV techniques. We describe the Copilot language and compiler, specifically designed for generating monitors for distributed, hard real-time systems. We also describe two case-studies in which we generated Copilot monitors in avionics systems.

  2. Adaptive Soa Stack-Based Business Process Monitoring Platform

    Directory of Open Access Journals (Sweden)

    Przemysław Dadel

    2014-01-01

    Full Text Available Executable business processes that formally describe company activities are well placed in the SOA environment as they allow for declarative organization of high-level system logic.However, for both technical and non-technical users, to fully benet from that element of abstractionappropriate business process monitoring systems are required and existing solutions remain unsatisfactory.The paper discusses the problem of business process monitoring in the context of the service orientation paradigm in order to propose an architectural solution and provide implementation of a system for business process monitoring that alleviates the shortcomings of the existing solutions.Various platforms are investigated to obtain a broader view of the monitoring problem and to gather functional and non-functional requirements. These requirements constitute input forthe further analysis and the system design. The monitoring software is then implemented and evaluated according to the specied criteria.An extensible business process monitoring system was designed and built on top of OSGiMM - a dynamic, event-driven, congurable communications layer that provides real-time monitoring capabilities for various types of resources. The system was tested against the stated functional requirements and its implementation provides a starting point for the further work.It is concluded that providing a uniform business process monitoring solution that satises a wide range of users and business process platform vendors is a dicult endeavor. It is furthermore reasoned that only an extensible, open-source, monitoring platform built on top of a scalablecommunication core has a chance to address all the stated and future requirements.

  3. A day in the life of a monitor!

    Directory of Open Access Journals (Sweden)

    Kunal Shah

    2012-01-01

    Full Text Available When at a site, the monitor will meet with the Study Coordinator, review the hospital medical records, use the internet database or paper to ′monitor′ their data versus their medical records, issue queries, check master files, count tablets or vials, provide the update to the doctor, and so on. When not traveling, the monitor will work in the office, printing letters, filing documents collected from sites, writing reports and follow-up letters, responding to e-mails, calling sites, to follow-up on the pending action items, in addition to calling sites not visited recently, attending study teleconferences, attending study and company training programs, reading standard operating procedures, completing excel spreadsheets or company specific software systems, and so on. The monitor is loaded with all these different types of work requirements and most importantly each and every task is important and time bound. Different skill sets are required for different tasks and the monitor plays different roles, while doing different tasks. This article enlists the tasks that are required to be done by the monitor, the different roles played by the monitor while doing these tasks, analyze which is the most important day for a monitor, what are the tasks performed during this day, and what knowledge and skills are required for performing these tasks.

  4. High-Level software requirements specification for the TWRS controlled baseline database system

    International Nuclear Information System (INIS)

    Spencer, S.G.

    1998-01-01

    This Software Requirements Specification (SRS) is an as-built document that presents the Tank Waste Remediation System (TWRS) Controlled Baseline Database (TCBD) in its current state. It was originally known as the Performance Measurement Control System (PMCS). Conversion to the new system name has not occurred within the current production system. Therefore, for simplicity, all references to TCBD are equivalent to PMCS references. This SRS will reference the PMCS designator from this point forward to capture the as-built SRS. This SRS is written at a high-level and is intended to provide the design basis for the PMCS. The PMCS was first released as the electronic data repository for cost, schedule, and technical administrative baseline information for the TAAS Program. During its initial development, the PMCS was accepted by the customer, TARS Business Management, with no formal documentation to capture the initial requirements

  5. Application of online chemistry monitoring programs and technology

    International Nuclear Information System (INIS)

    Perkins, D.; Choi, S.; Haas, C.

    2010-01-01

    To fully understand the impact of chemistry changes, several plant parameters must be considered and reviewed with actual chemistry analyses and compared to plant operating parameters. In some cases, this requires the ability to rapidly correlate plant operational data with laboratory and chemistry data. An effective online monitoring system should be able to: Integrate and extract online data from the plant laboratory and operating information from various plant data sources continuously; Interrogate and extract laboratory data from manually entered data on predefined frequencies; Interact with multiple laboratories in multiple locations; Evaluate data against plant limits (calculated or static) and provide personnel with action level or notification of plant exceeding limits; and, Provide the ability to evaluate against a standard and site specific set of calculations. The nuclear power industry continues to refine and apply new technologies in an effort to notify operators of changes in chemical conditions, calculate complex high temperature results, and monitor system performance. EPRI developed software specifically focused on plant chemistry program optimization using power operation and shutdown data applied with plant equipment. This software evolved into the family of software referred to as EPRI ChemWorks™ applications. As technology changed and improved, the application of online monitoring was essential for plant personnel working offsite. These changes in technology prompted EPRI to the development of SMART ChemWorks™ using the EPRI ChemWorks™ plant chemistry simulator and MULTEQ applications as the backbone for these simulations and calculations. SMART ChemWorks™ is an online monitoring system that queries plant databases and continuously monitors plant and chemistry parameters. The system uses a real-time intelligence engine to perform virtual sensing, identify normal and off-normal conditions and compare in-line instrument output to grab

  6. 30 CFR 250.457 - What equipment is required to monitor drilling fluids?

    Science.gov (United States)

    2010-07-01

    ... monitor the drilling fluid returns. The indicator may be located in the drilling fluid-logging compartment... 30 Mineral Resources 2 2010-07-01 2010-07-01 false What equipment is required to monitor drilling... OFFSHORE OIL AND GAS AND SULPHUR OPERATIONS IN THE OUTER CONTINENTAL SHELF Oil and Gas Drilling Operations...

  7. SU-E-J-211: Design and Study of In-House Software Based Respiratory Motion Monitoring, Controlling and Breath-Hold Device for Gated Radiotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Shanmugam, Senthilkumar [Madurai Medical College ' Govt. Rajaji Hospital, Madurai (India)

    2014-06-01

    Purpose: The purpose of this present work was to fabricate an in-house software based respiratory monitoring, controlling and breath-hold device using computer software programme which guides the patient to have uniform breath hold in response to request during the gated radiotherapy. Methods: The respiratory controlling device consists of a computer, inhouse software, video goggles, a highly sensitive sensor for measurement of distance, mounting systems, a camera, a respiratory signal device, a speaker and a visual indicator. The computer is used to display the respiratory movements of the patient with digital as well as analogue respiration indicators during the respiration cycle, to control, breath-hold and analyze the respiratory movement using indigenously developed software. Results: Studies were conducted with anthropomophic phantoms by simulating the respiratory motion on phantoms and recording the respective movements using the respiratory monitoring device. The results show good agreement between the simulated and measured movements. Further studies were conducted for 60 cancer patients with several types of cancers in the thoracic region. The respiratory movement cycles for each fraction of radiotherapy treatment were recorded and compared. Alarm indications are provided in the system to indicate when the patient breathing movement exceeds the threshold level. This will help the patient to maintain uniform breath hold during the radiotherapy treatment. Our preliminary clinical test results indicate that our device is highly reliable and able to maintain the uniform respiratory motion and breathe hold during the entire course of gated radiotherapy treatment. Conclusion: An indigenous respiratory monitoring device to guide the patient to have uniform breath hold device was fabricated. The alarm feature and the visual waveform indicator in the system guide the patient to have normal respiration. The signal from the device can be connected to the radiation

  8. SU-E-J-211: Design and Study of In-House Software Based Respiratory Motion Monitoring, Controlling and Breath-Hold Device for Gated Radiotherapy

    International Nuclear Information System (INIS)

    Shanmugam, Senthilkumar

    2014-01-01

    Purpose: The purpose of this present work was to fabricate an in-house software based respiratory monitoring, controlling and breath-hold device using computer software programme which guides the patient to have uniform breath hold in response to request during the gated radiotherapy. Methods: The respiratory controlling device consists of a computer, inhouse software, video goggles, a highly sensitive sensor for measurement of distance, mounting systems, a camera, a respiratory signal device, a speaker and a visual indicator. The computer is used to display the respiratory movements of the patient with digital as well as analogue respiration indicators during the respiration cycle, to control, breath-hold and analyze the respiratory movement using indigenously developed software. Results: Studies were conducted with anthropomophic phantoms by simulating the respiratory motion on phantoms and recording the respective movements using the respiratory monitoring device. The results show good agreement between the simulated and measured movements. Further studies were conducted for 60 cancer patients with several types of cancers in the thoracic region. The respiratory movement cycles for each fraction of radiotherapy treatment were recorded and compared. Alarm indications are provided in the system to indicate when the patient breathing movement exceeds the threshold level. This will help the patient to maintain uniform breath hold during the radiotherapy treatment. Our preliminary clinical test results indicate that our device is highly reliable and able to maintain the uniform respiratory motion and breathe hold during the entire course of gated radiotherapy treatment. Conclusion: An indigenous respiratory monitoring device to guide the patient to have uniform breath hold device was fabricated. The alarm feature and the visual waveform indicator in the system guide the patient to have normal respiration. The signal from the device can be connected to the radiation

  9. Virtual environment and computer-aided technologies used for system prototyping and requirements development

    Science.gov (United States)

    Logan, Cory; Maida, James; Goldsby, Michael; Clark, Jim; Wu, Liew; Prenger, Henk

    1993-01-01

    The Space Station Freedom (SSF) Data Management System (DMS) consists of distributed hardware and software which monitor and control the many onboard systems. Virtual environment and off-the-shelf computer technologies can be used at critical points in project development to aid in objectives and requirements development. Geometric models (images) coupled with off-the-shelf hardware and software technologies were used in The Space Station Mockup and Trainer Facility (SSMTF) Crew Operational Assessment Project. Rapid prototyping is shown to be a valuable tool for operational procedure and system hardware and software requirements development. The project objectives, hardware and software technologies used, data gained, current activities, future development and training objectives shall be discussed. The importance of defining prototyping objectives and staying focused while maintaining schedules are discussed along with project pitfalls.

  10. 5 CFR 575.212 - Internal monitoring requirements and revocation or suspension of authority.

    Science.gov (United States)

    2010-01-01

    ... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Internal monitoring requirements and... MANAGEMENT CIVIL SERVICE REGULATIONS RECRUITMENT, RELOCATION, AND RETENTION INCENTIVES; SUPERVISORY DIFFERENTIALS; AND EXTENDED ASSIGNMENT INCENTIVES Relocation Incentives § 575.212 Internal monitoring...

  11. 5 CFR 575.312 - Internal monitoring requirements and revocation or suspension of authority.

    Science.gov (United States)

    2010-01-01

    ... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Internal monitoring requirements and... MANAGEMENT CIVIL SERVICE REGULATIONS RECRUITMENT, RELOCATION, AND RETENTION INCENTIVES; SUPERVISORY DIFFERENTIALS; AND EXTENDED ASSIGNMENT INCENTIVES Retention Incentives § 575.312 Internal monitoring...

  12. Object oriented software for simulation and reconstruction of big alignment systems

    CERN Document Server

    Arce, P

    2003-01-01

    Modern high-energy physics experiments require tracking detectors to provide high precision under difficult working conditions (high magnetic field, gravity loads and temperature gradients). This is the reason why several of them are deciding to implement optical alignment systems to monitor the displacement of tracking elements in operation. To simulate and reconstruct optical alignment systems a general purpose software, named COCOA, has been developed, using the object oriented paradigm and software engineering techniques. Thanks to the big flexibility in its design, COCOA is able to reconstruct any optical system made of a combination of the following objects: laser, x-hair laser, incoherent source - pinhole, lens, mirror, plate splitter, cube splitter, optical square, rhomboid prism, 2D sensor, 1D sensor, distance-meter, tilt-meter, user-defined. COCOA was designed to satisfy the requirements of the CMS alignment system, which has several thousands of components. Sparse matrix techniques had been investi...

  13. Views on Software Testability

    OpenAIRE

    Shimeall, Timothy; Friedman, Michael; Chilenski, John; Voas, Jeffrey

    1994-01-01

    The field of testability is an active, well-established part of engineering of modern computer systems. However, only recently have technologies for software testability began to be developed. These technologies focus on accessing the aspects of software that improve or depreciate the ease of testing. As both the size of implemented software and the amount of effort required to test that software increase, so will the important of software testability technologies in influencing the softwa...

  14. Software for computerised analysis of cardiotocographic traces.

    Science.gov (United States)

    Romano, M; Bifulco, P; Ruffo, M; Improta, G; Clemente, F; Cesarelli, M

    2016-02-01

    Despite the widespread use of cardiotocography in foetal monitoring, the evaluation of foetal status suffers from a considerable inter and intra-observer variability. In order to overcome the main limitations of visual cardiotocographic assessment, computerised methods to analyse cardiotocographic recordings have been recently developed. In this study, a new software for automated analysis of foetal heart rate is presented. It allows an automatic procedure for measuring the most relevant parameters derivable from cardiotocographic traces. Simulated and real cardiotocographic traces were analysed to test software reliability. In artificial traces, we simulated a set number of events (accelerations, decelerations and contractions) to be recognised. In the case of real signals, instead, results of the computerised analysis were compared with the visual assessment performed by 18 expert clinicians and three performance indexes were computed to gain information about performances of the proposed software. The software showed preliminary performance we judged satisfactory in that the results matched completely the requirements, as proved by tests on artificial signals in which all simulated events were detected from the software. Performance indexes computed in comparison with obstetricians' evaluations are, on the contrary, not so satisfactory; in fact they led to obtain the following values of the statistical parameters: sensitivity equal to 93%, positive predictive value equal to 82% and accuracy equal to 77%. Very probably this arises from the high variability of trace annotation carried out by clinicians. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  15. Design and development of a run-time monitor for multi-core architectures in cloud computing.

    Science.gov (United States)

    Kang, Mikyung; Kang, Dong-In; Crago, Stephen P; Park, Gyung-Leen; Lee, Junghoon

    2011-01-01

    Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM) which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data.

  16. Design and Development of a Run-Time Monitor for Multi-Core Architectures in Cloud Computing

    Directory of Open Access Journals (Sweden)

    Junghoon Lee

    2011-03-01

    Full Text Available Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data.

  17. Towards Real-time, On-board, Hardware-Supported Sensor and Software Health Management for Unmanned Aerial Systems

    Science.gov (United States)

    Schumann, Johann; Rozier, Kristin Y.; Reinbacher, Thomas; Mengshoel, Ole J.; Mbaya, Timmy; Ippolito, Corey

    2013-01-01

    Unmanned aerial systems (UASs) can only be deployed if they can effectively complete their missions and respond to failures and uncertain environmental conditions while maintaining safety with respect to other aircraft as well as humans and property on the ground. In this paper, we design a real-time, on-board system health management (SHM) capability to continuously monitor sensors, software, and hardware components for detection and diagnosis of failures and violations of safety or performance rules during the flight of a UAS. Our approach to SHM is three-pronged, providing: (1) real-time monitoring of sensor and/or software signals; (2) signal analysis, preprocessing, and advanced on the- fly temporal and Bayesian probabilistic fault diagnosis; (3) an unobtrusive, lightweight, read-only, low-power realization using Field Programmable Gate Arrays (FPGAs) that avoids overburdening limited computing resources or costly re-certification of flight software due to instrumentation. Our implementation provides a novel approach of combining modular building blocks, integrating responsive runtime monitoring of temporal logic system safety requirements with model-based diagnosis and Bayesian network-based probabilistic analysis. We demonstrate this approach using actual data from the NASA Swift UAS, an experimental all-electric aircraft.

  18. SNM gamma-ray fingerprint monitor functional requirements and design specifications

    International Nuclear Information System (INIS)

    Bieber, A.M. Jr.; Kane, W.R.

    1994-07-01

    A number of DOE facilities need to perform confirmatory inventory measurements on items of special nuclear material (SNM). The DOE Office of Safeguards and Security (OSS) has tasked the Safeguards, Safety and Nonproliferation Division (SSN) of the Department of Advanced Technology at Brookhaven National Laboratory (BNL) to develop a high-resolution gamma-ray-spectroscopy-based instrument for performing confirmatory inventory measurements on such materials, a ''gamma-ray fingerprint monitor'' (GRFM). This document is a conceptual design for the SSN GRFM system. This conceptual design is based on previous experience with measurements of plutonium-bearing materials and comparison of gamma-ray spectrum features, not on actual tests of the procedures or hardware described. As a result, modifications may be necessary when actual prototype hardware and software are tested in realistic circumstances on actual materials of interest

  19. Software Quality Assurance Metrics

    Science.gov (United States)

    McRae, Kalindra A.

    2004-01-01

    Software Quality Assurance (SQA) is a planned and systematic set of activities that ensures conformance of software life cycle processes and products conform to requirements, standards and procedures. In software development, software quality means meeting requirements and a degree of excellence and refinement of a project or product. Software Quality is a set of attributes of a software product by which its quality is described and evaluated. The set of attributes includes functionality, reliability, usability, efficiency, maintainability, and portability. Software Metrics help us understand the technical process that is used to develop a product. The process is measured to improve it and the product is measured to increase quality throughout the life cycle of software. Software Metrics are measurements of the quality of software. Software is measured to indicate the quality of the product, to assess the productivity of the people who produce the product, to assess the benefits derived from new software engineering methods and tools, to form a baseline for estimation, and to help justify requests for new tools or additional training. Any part of the software development can be measured. If Software Metrics are implemented in software development, it can save time, money, and allow the organization to identify the caused of defects which have the greatest effect on software development. The summer of 2004, I worked with Cynthia Calhoun and Frank Robinson in the Software Assurance/Risk Management department. My task was to research and collect, compile, and analyze SQA Metrics that have been used in other projects that are not currently being used by the SA team and report them to the Software Assurance team to see if any metrics can be implemented in their software assurance life cycle process.

  20. Ontario Hydro experience in the identification and mitigation of potential failures in safety critical software systems

    International Nuclear Information System (INIS)

    Huget, R.G.; Viola, M.; Froebel, P.A.

    1995-01-01

    Ontario Hydro has had experience in designing and qualifying safety critical software used in the reactor shutdown systems of its nuclear generating stations. During software design, an analysis of system level hazards and potential hardware failure effects provide input to determining what safeguards will be needed. One form of safeguard, called software self checks, continually monitor the health of the computer on line. The design of self checks usually is a trade off between the amount of computing resources required, the software complexity, and the level of safeguarding provided. As part of the software verification activity, a software hazards analysis is performed, which identifiers any failure modes that could lead to the software causing an unsafe state, and which recommends changes to mitigate that potential. These recommendations may involve a re-structuring of the software to be more resistant to failure, or the introduction of other safeguarding measures. This paper discusses how Ontario Hydro has implemented these aspects of software design and verification into safety critical software used in reactor shutdown systems

  1. Verification of safety critical software

    International Nuclear Information System (INIS)

    Son, Ki Chang; Chun, Chong Son; Lee, Byeong Joo; Lee, Soon Sung; Lee, Byung Chai

    1996-01-01

    To assure quality of safety critical software, software should be developed in accordance with software development procedures and rigorous software verification and validation should be performed. Software verification is the formal act of reviewing, testing of checking, and documenting whether software components comply with the specified requirements for a particular stage of the development phase[1]. New software verification methodology was developed and was applied to the Shutdown System No. 1 and 2 (SDS1,2) for Wolsung 2,3 and 4 nuclear power plants by Korea Atomic Energy Research Institute(KAERI) and Atomic Energy of Canada Limited(AECL) in order to satisfy new regulation requirements of Atomic Energy Control Boars(AECB). Software verification methodology applied to SDS1 for Wolsung 2,3 and 4 project will be described in this paper. Some errors were found by this methodology during the software development for SDS1 and were corrected by software designer. Outputs from Wolsung 2,3 and 4 project have demonstrated that the use of this methodology results in a high quality, cost-effective product. 15 refs., 6 figs. (author)

  2. A cryogenic monitor system for the Liquid Argon Calorimeter in the SLD detector

    International Nuclear Information System (INIS)

    Fox, M.J.; Fox, J.D.

    1988-10-01

    This paper describes the monitoring electronics system design for the Liquid Argon Calorimeter (LAC) portion of the SLD detector. This system measures temperatures and liquid levels inside the LAC cryostat and transfers the results over a fiber-optic serial link to an external monitoring computer. System requirements, unique design constraints, and detailed analog, digital and software designs are presented. Fault tolerance and the requirement for a single design to work in several different operating environments are discussed. 4 refs., 3 figs., 1 tab

  3. Optimisation of the DQM software

    CERN Document Server

    Simetka, Vojtech

    2014-01-01

    The Data Quality Monitoring (DQM) Software proved to be a central tool in the CMS experiment. Its flexibility allowed its integration in several environments: Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release Validation, to constantly validate the functionality and the performance of the reconstruction soft- ware; in Monte Carlo productions [1, 2, 3]. The resources used by the different DQM software systems, both in terms of time and memory, are ever-growing. Having in mind the next to come LHC phases, the improvement of the perfor- mance of the DQM and Validation code will make possible the validation of extreme condition scenarios (events with high number of pileup vertices).

  4. The JOYO remote monitoring system

    International Nuclear Information System (INIS)

    Damico, Joseph P.; Hashimoto, Yu

    2000-01-01

    The evolution of the personal computer, operating systems and applications software and the Internet has brought drastic change and many benefits worldwide. Remote monitoring systems benefit from computer network and other modern software technologies. The availability of fast, inexpensive and secure communications enables new solutions for monitoring system applications. The JOYO Remote Monitoring System (RMS) utilizes computer network communications and modular software design to provide a distributed integrated solution for monitoring multiple storage locations. This paper describes the remote monitoring system installed at the JOYO Fast Reactor. The system combines sensors, software, and computer network technologies to create a powerful data collection, storage and dissemination capability. The RMS provides a flexible, scalable solution for a variety of applications. The RMS integrates a variety of state of the art technologies from several sources and serves as a test bed for cutting edge technologies that can be shared with outside users. This paper describes the system components and their operation and discusses system benefits. Current activities and future plants for the JOYO RMS will be discussed. (author)

  5. PRACTISE – Photo Rectification And ClassificaTIon SoftwarE (V.1.0

    Directory of Open Access Journals (Sweden)

    S. Härer

    2013-06-01

    Full Text Available Terrestrial photography is a cost-effective and easy-to-use method for measuring and monitoring spatially distributed land surface variables. It can be used to continuously investigate remote and often inaccessible terrain. We focus on the observation of snow cover patterns in high mountainous areas. The high temporal and spatial resolution of the photographs have various applications, for example validating spatially distributed snow hydrological models. However, the analysis of a photograph requires a preceding georectification of the digital camera image. To accelerate and simplify the analysis, we have developed the "Photo Rectification And ClassificaTIon SoftwarE" (PRACTISE that is available as a Matlab code. The routine requires a digital camera image, the camera location and its orientation, as well as a digital elevation model (DEM as input. If the viewing orientation and position of the camera are not precisely known, an optional optimisation routine using ground control points (GCPs helps to identify the missing parameters. PRACTISE also calculates a viewshed using the DEM and the camera position. The visible DEM pixels are utilised to georeference the photograph which is subsequently classified. The resulting georeferenced and classified image can be directly compared to other georeferenced data and can be used within any geoinformation system. The Matlab routine was tested using observations of the north-eastern slope of the Schneefernerkopf, Zugspitze, Germany. The results obtained show that PRACTISE is a fast and user-friendly tool, able to derive the microscale variability of snow cover extent in high alpine terrain, but can also easily be adapted to other land surface applications.

  6. Supporting Early Math--Rationales and Requirements for High Quality Software

    Science.gov (United States)

    Haake, Magnus; Husain, Layla; Gulz, Agneta

    2015-01-01

    There is substantial evidence that preschooler's performance in early math is highly correlated to math performance throughout school as well as academic skills in general. One way to help children attain early math skills is by using targeted educational software and the paper discusses potential gains of using such software to support early math…

  7. Field experience with remote monitoring

    Energy Technology Data Exchange (ETDEWEB)

    Desrosiers, A.E. [Bartlett Services, Inc., Plymouth, MA (United States)

    1995-03-01

    The Remote Monitoring System (RMS) is a combination of Merlin Gerin detection hardware, digital data communications hardware, and computer software from Bartlett Services, Inc. (BSI) that can improve the conduct of reactor plant operations in several areas. Using the RMS can reduce radiation exposures to radiation protection technicians (RPTs), reduce radiation exposures to plant maintenance and operations personnel, and reduce the time required to complete maintenance and inspections during outages. The number of temporary RPTs required during refueling outages can also be reduced. Data from use of the RMS at a two power plants are presented to illustrate these points.

  8. Test process for the safety-critical embedded software

    International Nuclear Information System (INIS)

    Sung, Ahyoung; Choi, Byoungju; Lee, Jangsoo

    2004-01-01

    Digitalization of nuclear Instrumentation and Control (I and C) system requires high reliability of not only hardware but also software. Verification and Validation (V and V) process is recommended for software reliability. But a more quantitative method is necessary such as software testing. Most of software in the nuclear I and C system is safety-critical embedded software. Safety-critical embedded software is specified, verified and developed according to V and V process. Hence two types of software testing techniques are necessary for the developed code. First, code-based software testing is required to examine the developed code. Second, after code-based software testing, software testing affected by hardware is required to reveal the interaction fault that may cause unexpected results. We call the testing of hardware's influence on software, an interaction testing. In case of safety-critical embedded software, it is also important to consider the interaction between hardware and software. Even if no faults are detected when testing either hardware or software alone, combining these components may lead to unexpected results due to the interaction. In this paper, we propose a software test process that embraces test levels, test techniques, required test tasks and documents for safety-critical embedded software. We apply the proposed test process to safety-critical embedded software as a case study, and show the effectiveness of it. (author)

  9. Paladin Software Support Lab

    Data.gov (United States)

    Federal Laboratory Consortium — The Paladin Software Support Environment (SSE) occupies 2,241 square-feet. It contains the hardware and software tools required to support the Paladin Automatic Fire...

  10. 48 CFR 12.212 - Computer software.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 1 2010-10-01 2010-10-01 false Computer software. 12.212... software. (a) Commercial computer software or commercial computer software documentation shall be acquired... required to— (1) Furnish technical information related to commercial computer software or commercial...

  11. An analysis software of tritium distribution in food and environmental water in China

    International Nuclear Information System (INIS)

    Li Wenhong; Xu Cuihua; Ren Tianshan; Deng Guilong

    2006-01-01

    Objective: The purpose of developing this analysis-software of tritium distribution in food and environmental water is to collect tritium monitoring data, to analyze the data, both automatically, statistically and graphically, and to study and share the data. Methods: Based on the data obtained before, analysis-software is wrote by using VC++. NET as tool software. The software first transfers data from EXCEL into a database. It has additive function of data-append, so operators can embody new monitoring data easily. Results: After turning the monitoring data saved as EXCEL file by original researchers into a database, people can easily access them. The software provides a tool of distributing-analysis of tritium. Conclusion: This software is a first attempt of data-analysis about tritium level in food and environmental water in China. Data achieving, searching and analyzing become easily and directly with the software. (authors)

  12. Towards an Ontology of Software

    OpenAIRE

    Wang, Xiaowei

    2016-01-01

    Software is permeating every aspect of our personal and social life. And yet, the cluster of concepts around the notion of software, such as the notions of a software product, software requirements, software specifications, are still poorly understood with no consensus on the horizon. For many, software is just code, something intangible best defined in contrast with hardware, but it is not particularly illuminating. This erroneous notion, software is just code, presents both in the ontology ...

  13. 75 FR 81126 - Revisions to Lead Ambient Air Monitoring Requirements

    Science.gov (United States)

    2010-12-27

    ... tons per year of lead is necessary to provide sufficient information about airborne lead levels near... Revisions to Lead Ambient Air Monitoring Requirements AGENCY: Environmental Protection Agency (EPA). ACTION...) that revised the primary and secondary National Ambient Air Quality Standards (NAAQS) for lead and...

  14. Dealing with distributed intelligence in monitoring and control systems

    International Nuclear Information System (INIS)

    McLaren, R.A.

    1981-01-01

    The Euorpean Hybrid Spectrometer is built up of many individual detectors, each having widely varying monitoring and control requirements. With the advent of cheap microprocessor systems a shift from the concept of a single monitoring and control computer of that of distributed intelligent controllers has been economically feasible. A detector designer can now thoroughly test and debug a complete monitoring and control system on a local, dedicated micro-computer, while during operation, the central computer can be relieved of many simple repetitive tasks. Rapidly, however, it has become obvious that the designers of these systems have to take into account the final operational environment and build into both the hardware and software, features allowing easy integration into a central monitoring and control chain. In addition, the problems of maintenance and enventual modification have to be taken into consideration early in the development. Examples of currently operational systems will be briefly described to demonstrate how a set of basic guidelines plus standardisation of hardware/software can minimise the problems of integration and maintenance. Based on practical experience gained in the European Hybrid Spectrometer, investigations are proceeding on various possible alternatives for future micro-computer based monitoring and control systems. (orig.)

  15. 40 CFR Table 11 to Subpart G of... - Wastewater-Inspection and Monitoring Requirements for Waste Management Units

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 9 2010-07-01 2010-07-01 false Wastewater-Inspection and Monitoring Requirements for Waste Management Units 11 Table 11 to Subpart G of Part 63 Protection of Environment... and Monitoring Requirements for Waste Management Units To comply with Inspection or monitoring...

  16. CMS Space Monitoring

    Science.gov (United States)

    Ratnikova, N.; Huang, C.-H.; Sanchez-Hernandez, A.; Wildish, T.; Zhang, X.

    2014-06-01

    During the first LHC run, CMS stored about one hundred petabytes of data. Storage accounting and monitoring help to meet the challenges of storage management, such as efficient space utilization, fair share between users and groups and resource planning. We present a newly developed CMS space monitoring system based on the storage metadata dumps produced at the sites. The information extracted from the storage dumps is aggregated and uploaded to a central database. A web based data service is provided to retrieve the information for a given time interval and a range of sites, so it can be further aggregated and presented in the desired format. The system has been designed based on the analysis of CMS monitoring requirements and experiences of the other LHC experiments. In this paper, we demonstrate how the existing software components of the CMS data placement system, PhEDEx, have been re-used, dramatically reducing the development effort.

  17. CMS Space Monitoring

    Energy Technology Data Exchange (ETDEWEB)

    Ratnikova, N. [Fermilab; Huang, C.-H. [Fermilab; Sanchez-Hernandez, A. [CINVESTAV, IPN; Wildish, T. [Princeton U.; Zhang, X. [Beijing, Inst. High Energy Phys.

    2014-01-01

    During the first LHC run, CMS stored about one hundred petabytes of data. Storage accounting and monitoring help to meet the challenges of storage management, such as efficient space utilization, fair share between users and groups and resource planning. We present a newly developed CMS space monitoring system based on the storage metadata dumps produced at the sites. The information extracted from the storage dumps is aggregated and uploaded to a central database. A web based data service is provided to retrieve the information for a given time interval and a range of sites, so it can be further aggregated and presented in the desired format. The system has been designed based on the analysis of CMS monitoring requirements and experiences of the other LHC experiments. In this paper, we demonstrate how the existing software components of the CMS data placement system, PhEDEx, have been re-used, dramatically reducing the development effort.

  18. Implementing Software Safety in the NASA Environment

    Science.gov (United States)

    Wetherholt, Martha S.; Radley, Charles F.

    1994-01-01

    Until recently, NASA did not consider allowing computers total control of flight systems. Human operators, via hardware, have constituted the ultimate safety control. In an attempt to reduce costs, NASA has come to rely more and more heavily on computers and software to control space missions. (For example. software is now planned to control most of the operational functions of the International Space Station.) Thus the need for systematic software safety programs has become crucial for mission success. Concurrent engineering principles dictate that safety should be designed into software up front, not tested into the software after the fact. 'Cost of Quality' studies have statistics and metrics to prove the value of building quality and safety into the development cycle. Unfortunately, most software engineers are not familiar with designing for safety, and most safety engineers are not software experts. Software written to specifications which have not been safety analyzed is a major source of computer related accidents. Safer software is achieved step by step throughout the system and software life cycle. It is a process that includes requirements definition, hazard analyses, formal software inspections, safety analyses, testing, and maintenance. The greatest emphasis is placed on clearly and completely defining system and software requirements, including safety and reliability requirements. Unfortunately, development and review of requirements are the weakest link in the process. While some of the more academic methods, e.g. mathematical models, may help bring about safer software, this paper proposes the use of currently approved software methodologies, and sound software and assurance practices to show how, to a large degree, safety can be designed into software from the start. NASA's approach today is to first conduct a preliminary system hazard analysis (PHA) during the concept and planning phase of a project. This determines the overall hazard potential of

  19. Software Product Manager: A Mechanism to manage software products in small and medium ISVs

    NARCIS (Netherlands)

    Katchow, R.; van de Weerd, I.; Brinkkemper, S.; Rooswinkel, A.

    2009-01-01

    In this paper, we present SP Manager as an innovative tool for managing software products in small and medium independent software vendors (ISVs). This tool incorporates the operational software product management (SPM) processes focused on requirements management and release planning. By using

  20. Expected requirements in support tool for software process improvement in SMEs

    OpenAIRE

    Muñoz Mata, Mirna; Mejía Miranda, Jezreel; Amescua Seco, Antonio; Calvo-Manzano Villalón, José Antonio; Cuevas Agustín, Gonzalo; San Feliu Gilabert, Tomás

    2012-01-01

    Nowadays being competitive is an important challenge for software development organizations. In order to achieve this, since last years, software process improvement has been an obvious and logical way. Unfortunately, even when many organizations are motivated to implement software process initiatives, not all know how best to do so, especially in Small and Medium Enterprises (SMEs) where due to its especial features, they have to be carefully in how to manage its resources to assure their ma...

  1. Self-learning health monitoring algorithm in composite structures

    Science.gov (United States)

    Grassia, Luigi; Iannone, Michele; Califano, America; D'Amore, Alberto

    2018-02-01

    The paper describes a system that it is able of monitoring the health state of a composite structure in real time. The hardware of the system consists of a wire of strain sensors connected to a control unit. The software of the system elaborates the strain data and in real time is able to detect the presence of an eventual damage of the structures monitored with the strain sensors. The algorithm requires as input only the strains of the monitored structured measured on real time, i.e. those strains coming from the deformations of the composite structure due to the working loads. The health monitoring system does not require any additional device to interrogate the structure as often used in the literature, instead it is based on a self-learning procedure. The strain data acquired when the structure is healthy are used to set up the correlations between the strain in different positions of structure by means of neural network. Once the correlations between the strains in different position have been set up, these correlations act as a fingerprint of the healthy structure. In case of damage the correlation between the strains in the position of the structure near the damage will change due to the change of the stiffness of the structure caused by the damage. The developed software is able to recognize the change of the transfer function between the strains and consequently is able to detect the damage.

  2. A Novel Software Platform Extending Advances in Monitoring Technologies to On-demand Decision Support

    Science.gov (United States)

    Ormerod, R.; Scholl, M.

    2017-12-01

    Rapid evolution is occurring in the monitoring and assessment of air emissions and their impacts. The development of next generation lower cost sensor technologies creates the potential for much more intensive and far-reaching monitoring networks that provide spatially rich data. While much attention at present is being directed at the types and performance characteristics of sensor technologies, it is important also that the full potential of rich data sources be realized. Parallel to sensor developments, software platforms to display and manage data in real time are increasingly common adjuncts to sensor networks. However, the full value of data can be realized by extending platform capabilities to include complex scientific functions that are integrated into an action-oriented management framework. Depending on the purpose and nature of a monitoring network, there will be a variety of potential uses of the data or its derivatives, for example: statistical analysis for policy development, event analysis, real-time issue management including emergency response and complaints, and predictive management. Moving these functions into an on-demand, optionally mobile, environment greatly increases the value and accessibility of the data. Increased interplay between monitoring data and decision-making in an operational environment is optimised by a system that is designed with equal weight on technical robustness and user experience. A system now being used by several regulatory agencies and a larger number of industries in the US, Latin America, Europe, Australia and Asia has been developed to provide a wide range of on-demand decision-support in addition to the basic data collection, display and management that most platforms offer. With stable multi-year operation, the platform, known as Envirosuite, is assisting organisations to both reduce operating costs and improve environmental performance. Some current examples of its application across a range of applications

  3. Risk monitor - a tool for operational safety assessment risk monitor - user's manual

    International Nuclear Information System (INIS)

    Hari Prasad, M.; Vinod, Gopika; Saraf, R.K.; Ghosh, A.K.

    2006-06-01

    Probabilistic Safety Assessment has become a key tool as on today to identify and understand Nuclear Power Plant vulnerabilities. As a result of the availability of these PSA studies, there is a desire to use them to enhance plant safety and to operate the nuclear stations in the most efficient manner. Risk Monitor is a PC based tool, which computes the real time safety level and assists plant personnel to manage day-to-day activities. Risk Monitor is a PC based user friendly software tool used for modification and re-analysis of a nuclear Power plant. Operation of Risk Monitor is based on PSA methods for assisting in day to day applications. Risk Monitoring programs can assess the risk profile and are used to optimize the operation of Nuclear Power Plants with respect to a minimum risk level over the operating time. This report presents the background activities of Risk Monitor, its application areas and the step by step procedure for the user.to interact with the software. This software can be used with the PSA model of any Nuclear Power Plant. (author)

  4. Software

    Energy Technology Data Exchange (ETDEWEB)

    Macedo, R.; Budd, G.; Ross, E.; Wells, P.

    2010-07-15

    The software section of this journal presented new software programs that have been developed to help in the exploration and development of hydrocarbon resources. Software provider IHS Inc. has made additions to its geological and engineering analysis software tool, IHS PETRA, a product used by geoscientists and engineers to visualize, analyze and manage well production, well log, drilling, reservoir, seismic and other related information. IHS PETRA also includes a directional well module and a decline curve analysis module to improve analysis capabilities in unconventional reservoirs. Petris Technology Inc. has developed a software to help manage the large volumes of data. PetrisWinds Enterprise (PWE) helps users find and manage wellbore data, including conventional wireline and MWD core data; analysis core photos and images; waveforms and NMR; and external files documentation. Ottawa-based Ambercore Software Inc. has been collaborating with Nexen on the Petroleum iQ software for steam assisted gravity drainage (SAGD) producers. Petroleum iQ integrates geology and geophysics data with engineering data in 3D and 4D. Calgary-based Envirosoft Corporation has developed a software that reduces the costly and time-consuming effort required to comply with Directive 39 of the Alberta Energy Resources Conservation Board. The product includes an emissions modelling software. Houston-based Seismic Micro-Technology (SMT) has developed the Kingdom software that features the latest in seismic interpretation. Holland-based Joa Oil and Gas and Calgary-based Computer Modelling Group have both supplied the petroleum industry with advanced reservoir simulation software that enables reservoir interpretation. The 2010 software survey included a guide to new software applications designed to facilitate petroleum exploration, drilling and production activities. Oil and gas producers can use the products for a range of functions, including reservoir characterization and accounting. In

  5. COMSY - A Software Tool for Aging and Plant Life Management

    International Nuclear Information System (INIS)

    Zander, Andre; Nopper, Helmut

    2012-01-01

    A Plant-wide and systematic Aging and Plant Life Management is essential for the safe operation and/or availability of nuclear power plants. The Aging Management (AM) has the objective to monitor and control degradation effects for safety relevant Systems, Structures and Components (SSCs) which may compromise safety functions of the plant. The Plant Life Management (PLM) methodology also includes aging surveillance for availability relevant SSCs. AM and PLM cover mechanical components, electrical and I and C systems and civil structures All Aging and Plant Life Management rules call for a comprehensive approach, requiring the systematic collection of various aging and safety relevant data on a plant-wide basis. This data needs to be serviced and periodically evaluated. Due to the complexity of the process, this activity needs to be supported by a qualified software tool for the management of aging relevant data and associated documents (approx. 30 000 SSCs). In order to support the power plant operators AREVA NP has developed the software tool COMSY. The COMSY software with its integrated AM modules enables the design and setup of a knowledge-based power plant model compatible to the requirements of international and national rules (e.g. IAEA Safety Guide NS-G-2.12, KTA 1403). In this process, a key task is to identify and monitor degradation mechanisms. For this purpose the COMSY tool provides prognosis and trending functions, which are based on more than 30 years of experience in the evaluation of degradation effects and numerous experimental studies. Since 1998 COMSY has been applied successfully in more than fifty reactor units in this field. The current version 3.0 was revised completely and offers additional AM functions. All aging-relevant component data are compiled and allocated via an integrated power plant model. Owing to existing interfaces to other software solutions and flexible import functions, COMSY is highly compatible with already existing data

  6. Models for composing software : an analysis of software composition and objects

    NARCIS (Netherlands)

    Bergmans, Lodewijk

    1999-01-01

    In this report, we investigate component-based software construction with a focus on composition. In particular we try to analyze the requirements and issues for components and software composition. As a means to understand this research area, we introduce a canonical model for representing

  7. Increasing the impact of usability work in software development

    DEFF Research Database (Denmark)

    Uldall-Espersen, Tobias; Frøkjær, Erik

    2006-01-01

    Usability, Case Study, Software Engineering, Software Quality, Organizational Impact, Usability Requirement Management, CHI 2007 workshop......Usability, Case Study, Software Engineering, Software Quality, Organizational Impact, Usability Requirement Management, CHI 2007 workshop...

  8. Assuring Software Reliability

    Science.gov (United States)

    2014-08-01

    technologies and processes to achieve a required level of confidence that software systems and services function in the intended manner. 1.3 Security Example...that took three high-voltage lines out of service and a software fail- ure (a race condition3) that disabled the computing service that notified the... service had failed. Instead of analyzing the details of the alarm server failure, the reviewers asked why the following software assurance claim had

  9. Strengthening Software Authentication with the ROSE Software Suite

    International Nuclear Information System (INIS)

    White, G

    2006-01-01

    Many recent nonproliferation and arms control software projects include a software authentication regime. These include U.S. Government-sponsored projects both in the United States and in the Russian Federation (RF). This trend toward requiring software authentication is only accelerating. Demonstrating assurance that software performs as expected without hidden ''backdoors'' is crucial to a project's success. In this context, ''authentication'' is defined as determining that a software package performs only its intended purpose and performs said purpose correctly and reliably over the planned duration of an agreement. In addition to visual inspections by knowledgeable computer scientists, automated tools are needed to highlight suspicious code constructs, both to aid visual inspection and to guide program development. While many commercial tools are available for portions of the authentication task, they are proprietary and not extensible. An open-source, extensible tool can be customized to the unique needs of each project (projects can have both common and custom rules to detect flaws and security holes). Any such extensible tool has to be based on a complete language compiler. ROSE is precisely such a compiler infrastructure developed within the Department of Energy (DOE) and targeted at the optimization of scientific applications and user-defined libraries within large-scale applications (typically applications of a million lines of code). ROSE is a robust, source-to-source analysis and optimization infrastructure currently addressing large, million-line DOE applications in C and C++ (handling the full C, C99, C++ languages and with current collaborations to support Fortran90). We propose to extend ROSE to address a number of security-specific requirements, and apply it to software authentication for nonproliferation and arms control projects

  10. PC based vibration monitoring system

    International Nuclear Information System (INIS)

    Jain, Sanjay K.; Roy, D.A.; Pithawa, C.K.; Patil, R.K.

    2004-01-01

    Health of large rotating machinery gets reflected in the vibration signature of the rotor and supporting structures and proper recording of these signals and their analysis can give a clear picture of the health of the machine. Using these data and their trending, it is possible to predict an impending trouble in the machine so that preventive action can be taken in time and catastrophic failure can be avoided. Continuous monitoring and analysis can give quick warning and enable operator to take preventive measures. Reactor Control Division, BARC is developing a PC based Vibration monitoring system for turbo generator machinery. The System can acquire 20 vibration signals at a rate of 5000 samples per second and also 15 process signals at a rate of 100 samples/ sec. The software for vibration monitoring system includes acquisition modules, analysis modules and Graphical User Interface module. The acquisition module involves initialization, setting of required parameters and acquiring the data from PC-based data acquisition cards. The acquired raw vibration data is then stored for analysis using various software packages. The display and analysis of acquired data is done in LabVIEW 7.0 where the data is displayed in time as well as frequency domain along with the RMS value of the signal. (author)

  11. Simulation software support (S3) system a software testing and debugging tool

    International Nuclear Information System (INIS)

    Burgess, D.C.; Mahjouri, F.S.

    1990-01-01

    The largest percentage of technical effort in the software development process is accounted for debugging and testing. It is not unusual for a software development organization to spend over 50% of the total project effort on testing. In the extreme, testing of human-rated software (e.g., nuclear reactor monitoring, training simulator) can cost three to five times as much as all other software engineering steps combined. The Simulation Software Support (S 3 ) System, developed by the Link-Miles Simulation Corporation is ideally suited for real-time simulation applications which involve a large database with models programmed in FORTRAN. This paper will focus on testing elements of the S 3 system. In this paper system support software utilities are provided which enable the loading and execution of modules in the development environment. These elements include the Linking/Loader (LLD) for dynamically linking program modules and loading them into memory and the interactive executive (IEXEC) for controlling the execution of the modules. Features of the Interactive Symbolic Debugger (SD) and the Real Time Executive (RTEXEC) to support the unit and integrated testing will be explored

  12. Toward objective software process information : experiences from a case study

    NARCIS (Netherlands)

    Samalikova, J.; Kusters, R.J.; Trienekens, J.J.M.; Weijters, A.J.M.M.; Siemons, P.

    2011-01-01

    A critical problem in software development is the monitoring, control and improvement in the processes of software developers. Software processes are often not explicitly modeled, and manuals to support the development work contain abstract guidelines and procedures. Consequently, there are huge

  13. Software Dependability and Safety Evaluations ESA's Initiative

    Science.gov (United States)

    Hernek, M.

    ESA has allocated funds for an initiative to evaluate Dependability and Safety methods of Software. The objectives of this initiative are; · More extensive validation of Safety and Dependability techniques for Software · Provide valuable results to improve the quality of the Software thus promoting the application of Dependability and Safety methods and techniques. ESA space systems are being developed according to defined PA requirement specifications. These requirements may be implemented through various design concepts, e.g. redundancy, diversity etc. varying from project to project. Analysis methods (FMECA. FTA, HA, etc) are frequently used during requirements analysis and design activities to assure the correct implementation of system PA requirements. The criticality level of failures, functions and systems is determined and by doing that the critical sub-systems are identified, on which dependability and safety techniques are to be applied during development. Proper performance of the software development requires the development of a technical specification for the products at the beginning of the life cycle. Such technical specification comprises both functional and non-functional requirements. These non-functional requirements address characteristics of the product such as quality, dependability, safety and maintainability. Software in space systems is more and more used in critical functions. Also the trend towards more frequent use of COTS and reusable components pose new difficulties in terms of assuring reliable and safe systems. Because of this, its dependability and safety must be carefully analysed. ESA identified and documented techniques, methods and procedures to ensure that software dependability and safety requirements are specified and taken into account during the design and development of a software system and to verify/validate that the implemented software systems comply with these requirements [R1].

  14. Workflow-Based Software Development Environment

    Science.gov (United States)

    Izygon, Michel E.

    2013-01-01

    The Software Developer's Assistant (SDA) helps software teams more efficiently and accurately conduct or execute software processes associated with NASA mission-critical software. SDA is a process enactment platform that guides software teams through project-specific standards, processes, and procedures. Software projects are decomposed into all of their required process steps or tasks, and each task is assigned to project personnel. SDA orchestrates the performance of work required to complete all process tasks in the correct sequence. The software then notifies team members when they may begin work on their assigned tasks and provides the tools, instructions, reference materials, and supportive artifacts that allow users to compliantly perform the work. A combination of technology components captures and enacts any software process use to support the software lifecycle. It creates an adaptive workflow environment that can be modified as needed. SDA achieves software process automation through a Business Process Management (BPM) approach to managing the software lifecycle for mission-critical projects. It contains five main parts: TieFlow (workflow engine), Business Rules (rules to alter process flow), Common Repository (storage for project artifacts, versions, history, schedules, etc.), SOA (interface to allow internal, GFE, or COTS tools integration), and the Web Portal Interface (collaborative web environment

  15. 48 CFR 227.7203-10 - Contractor identification and marking of computer software or computer software documentation to...

    Science.gov (United States)

    2010-10-01

    ... operation of the software to display a restrictive rights legend or other license notice; and (2) Requires a... and marking of computer software or computer software documentation to be furnished with restrictive... Rights in Computer Software and Computer Software Documentation 227.7203-10 Contractor identification and...

  16. AUTOMATED LOW-COST PHOTOGRAMMETRY FOR FLEXIBLE STRUCTURE MONITORING

    Directory of Open Access Journals (Sweden)

    C. H. Wang

    2012-07-01

    Full Text Available Structural monitoring requires instruments which can provide high precision and accuracy, reliable measurements at good temporal resolution and rapid processing speeds. Long-term campaigns and flexible structures are regarded as two of the most challenging subjects in monitoring engineering structures. Long-term monitoring in civil engineering is generally considered to be labourintensive and financially expensive and it can take significant effort to arrange the necessary human resources, transportation and equipment maintenance. When dealing with flexible structure monitoring, it is of paramount importance that any monitoring equipment used is able to carry out rapid sampling. Low cost, automated, photogrammetric techniques therefore have the potential to become routinely viable for monitoring non-rigid structures. This research aims to provide a photogrammetric solution for long-term flexible structural monitoring purposes. The automated approach was achieved using low-cost imaging devices (mobile phones to replace traditional image acquisition stations and substantially reduce the equipment costs. A self-programmed software package was developed to deal with the hardware-software integration and system operation. In order to evaluate the performance of this low-cost monitoring system, a shaking table experiment was undertaken. Different network configurations and target sizes were used to determine the best configuration. A large quantity of image data was captured by four DSLR cameras and four mobile phone cameras respectively. These image data were processed using photogrammetric techniques to calculate the final results for the system evaluation.

  17. Developments in radioactive scrap monitoring

    International Nuclear Information System (INIS)

    Bellian, J.G.

    1997-01-01

    Over the past ten years there have been major developments in radiation monitoring systems used for detecting shielded radioactive sources in scrap metal. The extent of the problem and industry's awareness of the problem have both grown significantly during that time. The multimillion dollar expenses associated with decontamination after a source passes into the melt and the potential health hazard to employees and the public have added further impetus to the development of monitoring systems. Early attempts at scrap monitoring could detect some radiation, but testing with real life situations showed them to be virtually incapable of detecting shielded sources of radioactivity in incoming vehicles. More sophisticated detector technology and the development of advanced software made useful by more powerful microprocessors led to successive generations of monitoring systems with order-of-magnitude improvement in detection capability. The next generation includes larger detectors and more complex algorithms offering further improvement in truck and rail car monitoring. Complete solutions require monitoring at additional locations within the site, such as the charge bucket and conveyor lines, and at the scrap processor's site

  18. An assessment of monitoring requirements and costs of 'Reduced Emissions from Deforestation and Degradation'

    Directory of Open Access Journals (Sweden)

    McCallum Ian

    2009-08-01

    Full Text Available Abstract Background Negotiations on a future climate policy framework addressing Reduced Emissions from Deforestation and Degradation (REDD are ongoing. Regardless of how such a framework will be designed, many technical solutions of estimating forest cover and forest carbon stock change exist to support policy in monitoring and accounting. These technologies typically combine remotely sensed data with ground-based inventories. In this article we assess the costs of monitoring REDD based on available technologies and requirements associated with key elements of REDD policy. Results We find that the design of a REDD policy framework (and specifically its rules can have a significant impact on monitoring costs. Costs may vary from 0.5 to 550 US$ per square kilometre depending on the required precision of carbon stock and area change detection. Moreover, they follow economies of scale, i.e. single country or project solutions will face relatively higher monitoring costs. Conclusion Although monitoring costs are relatively small compared to other cost items within a REDD system, they should be shared not only among countries but also among sectors, because an integrated monitoring system would have multiple benefits for non-REDD management. Overcoming initialization costs and unequal access to monitoring technologies is crucial for implementation of an integrated monitoring system, and demands for international cooperation.

  19. What Counts in Software Process?

    DEFF Research Database (Denmark)

    Cohn, Marisa

    2009-01-01

    and conversations in negotiating between prescriptions from a model and the contingencies that arise in an enactment. A qualitative field study at two Agile software development companies was conducted to investigate the role of artifacts in the software development work and the relationship between these artifacts...... and the Software Process. Documentation of software requirements is a major concern among software developers and software researchers. Agile software development denotes a different relationship to documentation, one that warrants investigation. Empirical findings are presented which suggest a new understanding...

  20. Methods of Run-Time Error Detection in Distributed Process Control Software

    DEFF Research Database (Denmark)

    Drejer, N.

    In this thesis, methods of run-time error detection in application software for distributed process control is designed. The error detection is based upon a monitoring approach in which application software is monitored by system software during the entire execution. The thesis includes definition...... and constraint evaluation is designed for the modt interesting error types. These include: a) semantical errors in data communicated between application tasks; b) errors in the execution of application tasks; and c) errors in the timing of distributed events emitted by the application software. The design...... of error detection methods includes a high level software specification. this has the purpose of illustrating that the designed can be used in practice....

  1. Software Engineering Reviews and Audits

    CERN Document Server

    Summers, Boyd L

    2011-01-01

    Accurate software engineering reviews and audits have become essential to the success of software companies and military and aerospace programs. These reviews and audits define the framework and specific requirements for verifying software development efforts. Authored by an industry professional with three decades of experience, Software Engineering Reviews and Audits offers authoritative guidance for conducting and performing software first article inspections, and functional and physical configuration software audits. It prepares readers to answer common questions for conducting and perform

  2. Quality assurance for software important to safety

    International Nuclear Information System (INIS)

    2000-01-01

    Software applications play an increasingly relevant role in nuclear power plant systems. This is particularly true of software important to safety used in both: calculations for the design, testing and analysis of nuclear reactor systems (design, engineering and analysis software); and monitoring, control and safety functions as an integral part of the reactor systems (monitoring, control and safety system software). Computer technology is advancing at a fast pace, offering new possibilities in nuclear reactor design, construction, commissioning, operation, maintenance and decommissioning. These advances also present new issues which must be considered both by the utility and by the regulatory organization. Refurbishment of ageing instrumentation and control systems in nuclear power plants and new safety related application areas have emerged, with direct (e.g. interfaces with safety systems) and indirect (e.g. operator intervention) implications for safety. Currently, there exist several international standards and guides on quality assurance for software important to safety. However, none of the existing documents provides comprehensive guidance to the developer, manager and regulator during all phases of the software life-cycle. The present publication was developed taking into account the large amount of available documentation, the rapid development of software systems and the need for updated guidance on h ow to do it . It provides information and guidance for defining and implementing quality assurance programmes covering the entire life-cycle of software important to safety. Expected users are managers, performers and assessors from nuclear utilities, regulatory bodies, suppliers and technical support organizations involved with the development and use of software applied in nuclear power plants

  3. Expert System Software Assistant for Payload Operations

    Science.gov (United States)

    Rogers, Mark N.

    1997-01-01

    The broad objective of this expert system software based application was to demonstrate the enhancements and cost savings that can be achieved through expert system software utilization in a spacecraft ground control center. Spacelab provided a valuable proving ground for this advanced software technology; a technology that will be exploited and expanded for future ISS operations. Our specific focus was on demonstrating payload cadre command and control efficiency improvements through the use of "smart" software which monitors flight telemetry, provides enhanced schematic-based data visualization, and performs advanced engineering data analysis.

  4. Mechatronics in design of monitoring and diagnostic systems

    Energy Technology Data Exchange (ETDEWEB)

    Uhl, T.; Barszcz, T. [Univ. of Mining and Metallurgy, Krakow (Poland); Hanc, A. [Energocontrol Ltd., Krakow (Poland)

    2003-07-01

    Nowadays development of computer engineering in area of hardware and software gives new possibilities of monitoring and diagnostics system design. The paper presents analysis of new possible solutions for design of monitoring and diagnostic systems including; smart sensor design, modular software design and communication modules. New concept of monitoring system based on home page server solution (nano-server) is presented. Smart sensor design concept with embedded hardware for diagnostic application is shown. New software concept for monitoring and diagnostics automation and examples of applications of new design for condition monitoring based on proposed solution are carefully discussed. (orig.)

  5. Radiation dose monitoring in the clinical routine

    Energy Technology Data Exchange (ETDEWEB)

    Guberina, Nika [UK Essen (Germany). Radiology

    2017-04-15

    Here we describe the first clinical experiences regarding the use of an automated radiation dose management software to monitor the radiation dose of patients during routine examinations. Many software solutions for monitoring radiation dose have emerged in the last decade. The continuous progress in radiological techniques, new scan features, scanner generations and protocols are the primary challenge for radiation dose monitoring software systems. To simulate valid dose calculations, radiation dose monitoring systems have to follow current trends and stay constantly up-to-date. The dose management software is connected to all devices at our institute and conducts automatic data acquisition and radiation dose calculation. The system incorporates 18 virtual phantoms based on the Cristy phantom family, estimating doses in newborns to adults. Dose calculation relies on a Monte Carlo simulation engine. Our first practical experiences demonstrate that the software is capable of dose estimation in the clinical routine. Its implementation and use have some limitations that can be overcome. The software is promising and allows assessment of radiation doses, like organ and effective doses according to ICRP 60 and ICRP 103, patient radiation dose history and cumulative radiation doses. Furthermore, we are able to determine local diagnostic reference doses. The radiation dose monitoring software systems can facilitate networking between hospitals and radiological departments, thus refining radiation doses and implementing reference doses at substantially lower levels.

  6. Computer systems and software description for Standard-E+ Hydrogen Monitoring System (SHMS-E+)

    International Nuclear Information System (INIS)

    Tate, D.D.

    1997-01-01

    The primary function of the Standard-E+ Hydrogen Monitoring System (SHMS-E+) is to determine tank vapor space gas composition and gas release rate, and to detect gas release events. Characterization of the gas composition is needed for safety analyses. The lower flammability limit, as well as the peak burn temperature and pressure, are dependent upon the gas composition. If there is little or no knowledge about the gas composition, safety analyses utilize compositions that yield the worst case in a deflagration or detonation. Knowledge of the true composition could lead to reductions in the assumptions and therefore there may be a potential for a reduction in controls and work restrictions. Also, knowledge of the actual composition will be required information for the analysis that is needed to remove tanks from the Watch List. Similarly, the rate of generation and release of gases is required information for performing safety analyses, developing controls, designing equipment, and closing safety issues. This report outlines the computer system design layout description for the Standard-E+ Hydrogen Monitoring System

  7. Towards a Global Monitoring System for CMS Computing Operations

    Energy Technology Data Exchange (ETDEWEB)

    Bauerdick, L. A.T. [Fermilab; Sciaba, Andrea [CERN

    2012-01-01

    The operation of the CMS computing system requires a complex monitoring system to cover all its aspects: central services, databases, the distributed computing infrastructure, production and analysis workflows, the global overview of the CMS computing activities and the related historical information. Several tools are available to provide this information, developed both inside and outside of the collaboration and often used in common with other experiments. Despite the fact that the current monitoring allowed CMS to successfully perform its computing operations, an evolution of the system is clearly required, to adapt to the recent changes in the data and workload management tools and models and to address some shortcomings that make its usage less than optimal. Therefore, a recent and ongoing coordinated effort was started in CMS, aiming at improving the entire monitoring system by identifying its weaknesses and the new requirements from the stakeholders, rationalise and streamline existing components and drive future software development. This contribution gives a complete overview of the CMS monitoring system and a description of all the recent activities that have been started with the goal of providing a more integrated, modern and functional global monitoring system for computing operations.

  8. Patterns for Parallel Software Design

    CERN Document Server

    Ortega-Arjona, Jorge Luis

    2010-01-01

    Essential reading to understand patterns for parallel programming Software patterns have revolutionized the way we think about how software is designed, built, and documented, and the design of parallel software requires you to consider other particular design aspects and special skills. From clusters to supercomputers, success heavily depends on the design skills of software developers. Patterns for Parallel Software Design presents a pattern-oriented software architecture approach to parallel software design. This approach is not a design method in the classic sense, but a new way of managin

  9. Thermal performance monitoring and optimisation

    International Nuclear Information System (INIS)

    Sunde, Svein; Berg; Oeyvind

    1998-01-01

    Monitoring of the thermal efficiency of nuclear power plants is expected to become increasingly important as energy-market liberalisation exposes plants to increasing availability requirements and fiercer competition. The general goal in thermal performance monitoring is straightforward: to maximise the ratio of profit to cost under the constraints of safe operation. One may perceive this goal to be pursued in two ways, one oriented towards fault detection and cost-optimal predictive maintenance, and another determined at optimising target values of parameters in response to any component degradation detected, changes in ambient conditions, or the like. Annual savings associated with effective thermal-performance monitoring are expected to be in the order of $ 100 000 for power plants of representative size. A literature review shows that a number of computer systems for thermal-performance monitoring exists, either as prototypes or commercially available. The characteristics and needs of power plants may vary widely, however, and decisions concerning the exact scope, content and configuration of a thermal-performance monitor may well follow a heuristic approach. Furthermore, re-use of existing software modules may be desirable. Therefore, we suggest here the design of a flexible workbench for easy assembly of an experimental thermal-performance monitor at the Halden Project. The suggested design draws heavily on our extended experience in implementing control-room systems featured by assets like high levels of customisation, flexibility in configuration and modularity in structure, and on a number of relevant adjoining activities. The design includes a multi-computer communication system and a graphical user's interface, and aims at a system adaptable to any combination of in-house or end user's modules, as well as commercially available software. (author)

  10. Improving software requirements specification for safety-related systems using the fault tree developed by an object-based method

    International Nuclear Information System (INIS)

    Cepin, M.; Mavko, B.

    1998-01-01

    A modification of the fault tree analysis is presented. The new fault tree integrates structural and behavioral models of a system. Information on the system structure is captured in the name of each gate and basic event of the fault tree. Information on the system behavior is captured in their description. Behavior is expressed using the axiomatic notation based on first order predicate logic. The new fault tree is a useful model for analysis and improvement of software requirements specification. The benefit of such improvements is reduced probability of failures in specification, which in turn results in increased reliability of the software.(author)

  11. Software design of a general purpose data acquisition and control executive

    International Nuclear Information System (INIS)

    Labiak, W.G.; Minor, E.G.

    1981-01-01

    The software design of an executive which performs general purpose data acquisition, monitoring, and control is presented. The executive runs on a memory-based mini or micro-computer and communicates with a disk-based computer where data analysis and display are done. The executive design stresses reliability and versatility, and has yielded software which can provide control and monitoring for widely different hardware systems. Applications of this software on two major fusion energy experiments at Lawrence Livermore National Laboratory will be described

  12. Software engineering methodologies and tools

    Science.gov (United States)

    Wilcox, Lawrence M.

    1993-01-01

    Over the years many engineering disciplines have developed, including chemical, electronic, etc. Common to all engineering disciplines is the use of rigor, models, metrics, and predefined methodologies. Recently, a new engineering discipline has appeared on the scene, called software engineering. For over thirty years computer software has been developed and the track record has not been good. Software development projects often miss schedules, are over budget, do not give the user what is wanted, and produce defects. One estimate is there are one to three defects per 1000 lines of deployed code. More and more systems are requiring larger and more complex software for support. As this requirement grows, the software development problems grow exponentially. It is believed that software quality can be improved by applying engineering principles. Another compelling reason to bring the engineering disciplines to software development is productivity. It has been estimated that productivity of producing software has only increased one to two percent a year in the last thirty years. Ironically, the computer and its software have contributed significantly to the industry-wide productivity, but computer professionals have done a poor job of using the computer to do their job. Engineering disciplines and methodologies are now emerging supported by software tools that address the problems of software development. This paper addresses some of the current software engineering methodologies as a backdrop for the general evaluation of computer assisted software engineering (CASE) tools from actual installation of and experimentation with some specific tools.

  13. A contemporary method for monitoring indoor radon and environmental conditions at a remote test site

    International Nuclear Information System (INIS)

    Renken, K.J.; Coursin, S.

    1996-01-01

    A state-of-the-art method for automatically monitoring indoor radon and environmental conditions at a remote test site is described. A Wisconsin home that exhibited elevated radon levels has been installed with automated PC-data acquisition system (PC-DAS) that includes: a laptop PC, a data acquisition cardcage, a commercial data acquisition software program plus sensors to measure radon gas concentrations, differential pressures, indoor air quality and meteorological conditions. The isolated PC-DAS is connected to a PC in a university laboratory via a modem and a communications software package. Experimental data is monitored and saved by the remote PC in real time and then automatically downloaded to the lab computer at selected intervals. An example of the formatted field results is presented and analysed. This documentation of the set-up, the off-the-shelf computer hardware and software, and the procedures should assist investigations requiring flexible remote long-term radon and environmental monitoring. (Author)

  14. Integrated modeling of software cost and quality

    International Nuclear Information System (INIS)

    Rone, K.Y.; Olson, K.M.

    1994-01-01

    In modeling the cost and quality of software systems, the relationship between cost and quality must be considered. This explicit relationship is dictated by the criticality of the software being developed. The balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and the developers with respect to the processes being employed

  15. Requirements to micro-unmanned aircraft systems in civil protection and environmental monitoring

    Energy Technology Data Exchange (ETDEWEB)

    Fischer-Stabel, Peter; Hardt, Christopher [Univ. of Applied Sciences Trier, Birkenfeld (Germany). Dept. of Environmental Planning

    2013-07-01

    Especially in application fields such as environmental monitoring or in the field of information and operations management with technical or natural disasters, increased demands on communication and sensor technology to micro unmanned aircraft systems (UAS) are given. These are currently covered by the system manufacturers, however inadequately. The use case of wildlife monitoring with micro UAS comes with some special requirements and problems, addressed in this paper. (orig.)

  16. The Basic Requirements for Personnel Monitoring

    International Nuclear Information System (INIS)

    1965-01-01

    The code of practice presented here as part of the Agency's Safety Standards deals with the problem of personnel monitoring in establishments in which, or in part of which, work is primarily with radiation sources. It is a development of the fundamental requirements on the subject contained in the Agency's Basic Safety Standards. The code has been prepared by the Secretariat of the Agency on the basis of the work of a panel of experts drawn from a number of Member States. In addition, representatives of several interested international organizations participated in the work of the panel. The Board of Governors of the Agency considered this code of practice in September 1964 and authorized its application to Agency and Agency-assisted operations and recommended to Member States that it be taken into account in the formulation of national regulations or recommendations.

  17. Anticipating requirements changes-using futurology in requirements elicitation

    OpenAIRE

    Pimentel, João Henrique; Santos, Emanuel; Castro, Jaelson; Franch Gutiérrez, Javier

    2012-01-01

    It is well known that requirements changes in a later phase of software developments is a major source of software defects and costs. Thus, the need of techniques to control or reduce the amount of changes during software development projects. The authors advocate the use of foresight methods as a valuable input to requirements elicitation, with the potential to decrease the number of changes that would be required after deployment, by anticipating them. In this paper, the authors define a pr...

  18. Software architecture evolution

    DEFF Research Database (Denmark)

    Barais, Olivier; Le Meur, Anne-Francoise; Duchien, Laurence

    2008-01-01

    Software architectures must frequently evolve to cope with changing requirements, and this evolution often implies integrating new concerns. Unfortunately, when the new concerns are crosscutting, existing architecture description languages provide little or no support for this kind of evolution....... The software architect must modify multiple elements of the architecture manually, which risks introducing inconsistencies. This chapter provides an overview, comparison and detailed treatment of the various state-of-the-art approaches to describing and evolving software architectures. Furthermore, we discuss...... one particular framework named Tran SAT, which addresses the above problems of software architecture evolution. Tran SAT provides a new element in the software architecture descriptions language, called an architectural aspect, for describing new concerns and their integration into an existing...

  19. Essential software architecture

    CERN Document Server

    Gorton, Ian

    2011-01-01

    Job titles like ""Technical Architect"" and ""Chief Architect"" nowadays abound in software industry, yet many people suspect that ""architecture"" is one of the most overused and least understood terms in professional software development. Gorton's book tries to resolve this dilemma. It concisely describes the essential elements of knowledge and key skills required to be a software architect. The explanations encompass the essentials of architecture thinking, practices, and supporting technologies. They range from a general understanding of structure and quality attributes through technical i

  20. The ATLAS Trigger Simulation with Legacy Software

    CERN Document Server

    Bernius, Catrin; The ATLAS collaboration

    2017-01-01

    Physics analyses at the LHC require accurate simulations of the detector response and the event selection processes, generally done with the most recent software releases. The trigger response simulation is crucial for determination of overall selection efficiencies and signal sensitivities and should be done with the same software release with which data were recorded. This requires potentially running with software dating many years back, the so-called legacy software. Therefore having a strategy for running legacy software in a modern environment becomes essential when data simulated for past years start to present a sizeable fraction of the total. The requirements and possibilities for such a simulation scheme within the ATLAS software framework were examined and a proof-of-concept simulation chain has been successfully implemented. One of the greatest challenges was the choice of a data format which promises long term compatibility with old and new software releases. Over the time periods envisaged, data...

  1. A measurement system for large, complex software programs

    Science.gov (United States)

    Rone, Kyle Y.; Olson, Kitty M.; Davis, Nathan E.

    1994-01-01

    This paper describes measurement systems required to forecast, measure, and control activities for large, complex software development and support programs. Initial software cost and quality analysis provides the foundation for meaningful management decisions as a project evolves. In modeling the cost and quality of software systems, the relationship between the functionality, quality, cost, and schedule of the product must be considered. This explicit relationship is dictated by the criticality of the software being developed. This balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and developers with respect to the processes being employed.

  2. Integrated photovoltaic (PV) monitoring system

    Science.gov (United States)

    Mahinder Singh, Balbir Singh; Husain, NurSyahidah; Mohamed, Norani Muti

    2012-09-01

    The main aim of this research work is to design an accurate and reliable monitoring system to be integrated with solar electricity generating system. The performance monitoring system is required to ensure that the PVEGS is operating at an optimum level. The PV monitoring system is able to measure all the important parameters that determine an optimum performance. The measured values are recorded continuously, as the data acquisition system is connected to a computer, and data is stored at fixed intervals. The data can be locally used and can also be transmitted via internet. The data that appears directly on the local monitoring system is displayed via graphical user interface that was created by using Visual basic and Apache software was used for data transmission The accuracy and reliability of the developed monitoring system was tested against the data that captured simultaneously by using a standard power quality analyzer device. The high correlation which is 97% values indicates the level of accuracy of the monitoring system. The aim of leveraging on a system for continuous monitoring system is achieved, both locally, and can be viewed simultaneously at a remote system.

  3. Some remarks on software reliability

    International Nuclear Information System (INIS)

    Gonzalez Hernando, J.; Sanchez Izquierdo, J.

    1978-01-01

    Trend in modern NPPCI is toward a broad use of programmable elements. Some aspects concerning present status of programmable digital systems reliability are reported. Basic differences between software and hardware concept require a specific approach in all the reliability topics concerning software systems. The software reliability theory was initialy developed upon hardware models analogies. At present this approach is changing and specific models are being developed. The growing use of programmable systems necessitates emphasizing the importance of more adequate regulatory requirements to include this technology in NPPCI. (author)

  4. TWRS engineering bibliography software listing

    International Nuclear Information System (INIS)

    Husa, E.I.

    1995-01-01

    This document contains the computer software listing for Engineering Bibliography software, developed by E. Ivar Husa. This software is in the working prototype stage of development. The code has not been tested to requirements. TWRS Engineering created this software for engineers to share bibliographic references across the Hanford site network (HLAN). This software is intended to store several hundred to several thousand references (a compendium with limited range). Coded changes are needed to support the larger number of references

  5. Integrated software system for improving medical equipment management.

    Science.gov (United States)

    Bliznakov, Z; Pappous, G; Bliznakova, K; Pallikarakis, N

    2003-01-01

    The evolution of biomedical technology has led to an extraordinary use of medical devices in health care delivery. During the last decade, clinical engineering departments (CEDs) turned toward computerization and application of specific software systems for medical equipment management in order to improve their services and monitor outcomes. Recently, much emphasis has been given to patient safety. Through its Medical Device Directives, the European Union has required all member nations to use a vigilance system to prevent the reoccurrence of adverse events that could lead to injuries or death of patients or personnel as a result of equipment malfunction or improper use. The World Health Organization also has made this issue a high priority and has prepared a number of actions and recommendations. In the present workplace, a new integrated, Windows-oriented system is proposed, addressing all tasks of CEDs but also offering a global approach to their management needs, including vigilance. The system architecture is based on a star model, consisting of a central core module and peripheral units. Its development has been based on the integration of 3 software modules, each one addressing specific predefined tasks. The main features of this system include equipment acquisition and replacement management, inventory archiving and monitoring, follow up on scheduled maintenance, corrective maintenance, user training, data analysis, and reports. It also incorporates vigilance monitoring and information exchange for adverse events, together with a specific application for quality-control procedures. The system offers clinical engineers the ability to monitor and evaluate the quality and cost-effectiveness of the service provided by means of quality and cost indicators. Particular emphasis has been placed on the use of harmonized standards with regard to medical device nomenclature and classification. The system's practical applications have been demonstrated through a pilot

  6. Intelligent software solution for reliable high efficiency/low false alarm border monitoring

    International Nuclear Information System (INIS)

    Rieck, W.; Iwatschenko, M.

    2001-01-01

    Full text: Radioactivity Monitoring at border stations requires detection systems that are reliably operating under special conditions such as: different types and shapes of vehicles; different velocities; stop and go traffic. ESM has developed a solution that achieves under all such conditions the lowest possible detection limit and avoids false alarms generated by naturally occurring radioactive material (NORM). NBR (Natural Background Reduction) data evaluation - One of the main reasons for the success of the ESM gate monitors is the unique and proprietary NBR-technology of instantaneous discrimination of artificial and natural gamma radiation using large area plastic scintillators. Thus the FHT 1388 gate monitors show 2 unique features: Possible setting of different alarm levels for NORM and artificial gamma sources; Self adjusting compensation of the background shielding of the truck in respect to the detection of artificial sources. Both properties are a preposition for the highly sensitive detection of artificial gamma sources. While at scrap yards and steel mills usually all radioactivity (including NORM) must be detected, the main object of interest in respect to the measuring task at border stations, airports or harbours is clearly the detection of even very small signals of artificial radioactivity. The reliable rejection of the influence of natural radioactivity is of special importance in the case of detection of illicit trafficking, since construction material, fertilisers or soil often lead to much higher detector signals than the alarming levels for dangerous sources of interest. Beside the varying content of natural radioactivity in the load of a truck, different loads and trucks show different influence on the reduction of the ambient radiation due to the passing vehicle. Thus software approaches assuming a specific reduction of the background count rate (regarding relative magnitude and shape) must fail when trucks of different shape and load

  7. Software Engineering Improvement Activities/Plan

    Science.gov (United States)

    2003-01-01

    bd Systems personnel accomplished the technical responsibilities for this reporting period, as planned. A close working relationship was maintained with personnel of the MSFC Avionics Department Software Group (ED14). Work accomplishments included development, evaluation, and enhancement of a software cost model, performing literature search and evaluation of software tools available for code analysis and requirements analysis, and participating in other relevant software engineering activities. Monthly reports were submitted. This support was provided to the Flight Software Group/ED 1 4 in accomplishing the software engineering improvement engineering activities of the Marshall Space Flight Center (MSFC) Software Engineering Improvement Plan.

  8. Taxonomía de riesgos de outsourcing de software Software outsourcing risk taxonomy

    Directory of Open Access Journals (Sweden)

    Gloria Piedad Gasca-Hurtado

    2013-04-01

    Full Text Available Para las organizaciones la estrategia de outsourcing de software (OS se ha constituido en una alternativa estratégica de reducción de costos y mejora de la calidad de software. Sin embargo, la complejidad de la gestión de este tipo de proyectos aumenta, debido a la relación de actores diferentes (cliente y proveedor, de la cual depende en gran medida el éxito de estos proyectos. Identificar, analizar y gestionar los riesgos que afectan este tipo de proyectos, son buenas prácticas que se pueden llevar a cabo para establecer indicadores de gestión claves para el éxito de los proyectos. Es por esta razón que el objetivo de este trabajo es diseñar una taxonomía de riesgos de OS, es decir, una estructura de clasificación jerárquica de riesgos basada en los proyectos de OS que facilite la identificación de riesgos, que contribuya a la realización de un análisis de riesgos ágil y preciso, para su posterior gestión a lo largo de todo el proceso de OS. La taxonomía que se propone ha sido validada en un caso de estudio correspondiente a un proyecto de compra de software preelaborado. Los resultados se muestran en la última parte de este trabajo y se analiza la forma como dichos resultados impacta el proyecto en el cual fue implementada la taxonomía y su utilidad en el refinamiento de los procesos organizacionales de la empresa, a partir de las lecciones aprendidas.In recent years, the software outsourcing (SO strategy for organizations has become a useful and important alternative, as well as SO projects. SO projects have many complexities due to several factors affecting management. These complexities and importance of SO require that successfull projects meet a parameter set: budget, cost, schedule, among others. This requirement can be met through the implementation of methods and techniques of risk management (RM. RM is a process to prevent problems, establish actions to follow, and monitor the project to ensure success. Likewise

  9. Software for Optimizing Quality Assurance of Other Software

    Science.gov (United States)

    Feather, Martin; Cornford, Steven; Menzies, Tim

    2004-01-01

    Software assurance is the planned and systematic set of activities that ensures that software processes and products conform to requirements, standards, and procedures. Examples of such activities are the following: code inspections, unit tests, design reviews, performance analyses, construction of traceability matrices, etc. In practice, software development projects have only limited resources (e.g., schedule, budget, and availability of personnel) to cover the entire development effort, of which assurance is but a part. Projects must therefore select judiciously from among the possible assurance activities. At its heart, this can be viewed as an optimization problem; namely, to determine the allocation of limited resources (time, money, and personnel) to minimize risk or, alternatively, to minimize the resources needed to reduce risk to an acceptable level. The end result of the work reported here is a means to optimize quality-assurance processes used in developing software.

  10. Software Reuse Within the Earth Science Community

    Science.gov (United States)

    Marshall, James J.; Olding, Steve; Wolfe, Robert E.; Delnore, Victor E.

    2006-01-01

    Scientific missions in the Earth sciences frequently require cost-effective, highly reliable, and easy-to-use software, which can be a challenge for software developers to provide. The NASA Earth Science Enterprise (ESE) spends a significant amount of resources developing software components and other software development artifacts that may also be of value if reused in other projects requiring similar functionality. In general, software reuse is often defined as utilizing existing software artifacts. Software reuse can improve productivity and quality while decreasing the cost of software development, as documented by case studies in the literature. Since large software systems are often the results of the integration of many smaller and sometimes reusable components, ensuring reusability of such software components becomes a necessity. Indeed, designing software components with reusability as a requirement can increase the software reuse potential within a community such as the NASA ESE community. The NASA Earth Science Data Systems (ESDS) Software Reuse Working Group is chartered to oversee the development of a process that will maximize the reuse potential of existing software components while recommending strategies for maximizing the reusability potential of yet-to-be-designed components. As part of this work, two surveys of the Earth science community were conducted. The first was performed in 2004 and distributed among government employees and contractors. A follow-up survey was performed in 2005 and distributed among a wider community, to include members of industry and academia. The surveys were designed to collect information on subjects such as the current software reuse practices of Earth science software developers, why they choose to reuse software, and what perceived barriers prevent them from reusing software. In this paper, we compare the results of these surveys, summarize the observed trends, and discuss the findings. The results are very

  11. Proposed radioactive liquid effluent monitoring requirements at the Savannah River Site

    International Nuclear Information System (INIS)

    Jannik, G.T.; Carlton, W.H.; Blunt, B.C.

    1994-01-01

    Clear regulatory guidance exists for structuring a radiological air monitoring program, however, there is no parallel guidance for radiological liquid monitoring. For Department of Energy (DOE) facilities, there are no existing applicable federal regulations, DOE orders, or DOE guidance documents that specify at what levels continuous monitoring, continuous sampling, or periodic confirmatory measurements of radioactive liquid effluents must be made. In order to bridge this gap and to technically justify and document liquid effluent monitoring decisions at DOE's Savannah River Site, Westinghouse Savannah River Company has proposed that a graded, dose-based approach be established, in conjunction with limits on facility radionuclide inventories, to determine the monitoring and sampling criteria to be applied at each potential liquid radioactive effluent point. The graded approach would be similar to--and a conservative extension of--the existing, agreed-upon SRS/EPA-IV airborne effluent monitoring approach documented in WSRC's NESHAP Quality Assurance Project Plan. The limits on facility radionuclide inventories are based on--and are a conservative extension of--the 10 CFR 834, 10 CFR 20, and SCR 61-63 annual limits on discharges to sanitary sewers. Used in conjunction with each other, the recommended source category criteria levels and facility radionuclide inventories would allow for the best utilization of resources and provide consistent, technically justifiable determinations of radioactive liquid effluent monitoring requirements

  12. 10 CFR 20.1502 - Conditions requiring individual monitoring of external and internal occupational dose.

    Science.gov (United States)

    2010-01-01

    ... external and internal occupational dose. Each licensee shall monitor exposures to radiation and radioactive... 10 Energy 1 2010-01-01 2010-01-01 false Conditions requiring individual monitoring of external and internal occupational dose. 20.1502 Section 20.1502 Energy NUCLEAR REGULATORY COMMISSION STANDARDS FOR...

  13. Awareness during anaesthesia for surgery requiring evoked potential monitoring: A pilot study

    Directory of Open Access Journals (Sweden)

    Pritish J Korula

    2017-01-01

    Full Text Available Background: Evoked potential monitoring such as somatosensory-evoked potential (SSEP or motor-evoked potential (MEP monitoring during surgical procedures in proximity to the spinal cord requires minimising the minimum alveolar concentrations (MACs below the anaesthetic concentrations normally required (1 MAC to prevent interference in amplitude and latency of evoked potentials. This could result in awareness. Our primary objective was to determine the incidence of awareness while administering low MAC inhalational anaesthetics for these unique procedures. The secondary objective was to assess the adequacy of our anaesthetic technique from neurophysiologist′s perspective. Methods: In this prospective observational pilot study, 61 American Society of Anesthesiologists 1 and 2 patients undergoing spinal surgery for whom intraoperative evoked potential monitoring was performed were included; during the maintenance phase, 0.7-0.8 MAC of isoflurane was targeted. We evaluated the intraoperative depth of anaesthesia using a bispectral (BIS index monitor as well as the patients response to surgical stimulus (PRST scoring system. Post-operatively, a modified Bruce questionnaire was used to verify awareness. The adequacy of evoked potential readings was also assessed. Results: Of the 61 patients, no patient had explicit awareness. Intraoperatively, 19 of 61 patients had a BIS value of above sixty at least once, during surgery. There was no correlation with PRST scoring and BIS during surgery. Fifty-four out of 61 patient′s evoked potential readings were deemed ′good′ or ′fair′ for the conduct of electrophysiological monitoring. Conclusions: This pilot study demonstrates that administering low MAC inhalational anaesthetics to facilitate evoked potential monitoring does not result in explicit awareness. However, larger studies are needed to verify this. The conduct of SSEP electrophysiological monitoring was satisfactory with the use of this

  14. The safety implications of emerging software paradigms

    International Nuclear Information System (INIS)

    Suski, G.J.; Persons, W.L.; Johnson, G.L.

    1994-10-01

    This paper addresses some of the emerging software paradigms that may be used in developing safety-critical software applications. Paradigms considered in this paper include knowledge-based systems, neural networks, genetic algorithms, and fuzzy systems. It presents one view of the software verification and validation activities that should be associated with each paradigm. The paper begins with a discussion of the historical evolution of software verification and validation. Next, a comparison is made between the verification and validation processes used for conventional and emerging software systems. Several verification and validation issues for the emerging paradigms are discussed and some specific research topics are identified. This work is relevant for monitoring and control at nuclear power plants

  15. KAERI software safety guideline for developing safety-critical software in digital instrumentation and control system of nuclear power plant

    International Nuclear Information System (INIS)

    Lee, Jang Soo; Kim, Jang Yeol; Eum, Heung Seop.

    1997-07-01

    Recently, the safety planning for safety-critical software systems is being recognized as the most important phase in the software life cycle, and being developed new regulatory positions and standards by the regulatory and the standardization organization. The requirements for software important to safety of nuclear reactor are described in such positions and standards. Most of them are describing mandatory requirements, what shall be done, for the safety-critical software. The developers of such a software. However, there have been a lot of controversial factors on whether the work practices satisfy the regulatory requirements, and to justify the safety of such a system developed by the work practices, between the licenser and the licensee. We believe it is caused by the reason that there is a gap between the mandatory requirements (What) and the work practices (How). We have developed a guidance to fill such gap, which can be useful for both licenser and licensee to conduct a justification of the safety in the planning phase of developing the software for nuclear reactor protection systems. (author). 67 refs., 13 tabs., 2 figs

  16. KAERI software safety guideline for developing safety-critical software in digital instrumentation and control system of nuclear power plant

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Jang Soo; Kim, Jang Yeol; Eum, Heung Seop

    1997-07-01

    Recently, the safety planning for safety-critical software systems is being recognized as the most important phase in the software life cycle, and being developed new regulatory positions and standards by the regulatory and the standardization organization. The requirements for software important to safety of nuclear reactor are described in such positions and standards. Most of them are describing mandatory requirements, what shall be done, for the safety-critical software. The developers of such a software. However, there have been a lot of controversial factors on whether the work practices satisfy the regulatory requirements, and to justify the safety of such a system developed by the work practices, between the licenser and the licensee. We believe it is caused by the reason that there is a gap between the mandatory requirements (What) and the work practices (How). We have developed a guidance to fill such gap, which can be useful for both licenser and licensee to conduct a justification of the safety in the planning phase of developing the software for nuclear reactor protection systems. (author). 67 refs., 13 tabs., 2 figs.

  17. The ATLAS online High Level Trigger framework experience reusing offline software components in the ATLAS trigger

    CERN Document Server

    Wiedenmann, W

    2009-01-01

    Event selection in the Atlas High Level Trigger is accomplished to a large extent by reusing software components and event selection algorithms developed and tested in an offline environment. Many of these offline software modules are not specifically designed to run in a heavily multi-threaded online data flow environment. The Atlas High Level Trigger (HLT) framework based on the Gaudi and Atlas Athena frameworks, forms the interface layer, which allows the execution of the HLT selection and monitoring code within the online run control and data flow software. While such an approach provides a unified environment for trigger event selection across all of Atlas, it also poses strict requirements on the reused software components in terms of performance, memory usage and stability. Experience of running the HLT selection software in the different environments and especially on large multi-node trigger farms has been gained in several commissioning periods using preloaded Monte Carlo events, in data taking peri...

  18. Safety-Critical Partitioned Software Architecture: A Partitioned Software Architecture for Robotic

    Science.gov (United States)

    Horvath, Greg; Chung, Seung H.; Cilloniz-Bicchi, Ferner

    2011-01-01

    The flight software on virtually every mission currently managed by JPL has several major flaws that make it vulnerable to potentially fatal software defects. Many of these problems can be addressed by recently developed partitioned operating systems (OS). JPL has avoided adopting a partitioned operating system on its flight missions, primarily because doing so would require significant changes in flight software design, and the risks associated with changes of that magnitude cannot be accepted by an active flight project. The choice of a partitioned OS can have a dramatic effect on the overall system and software architecture, allowing for realization of benefits far beyond the concerns typically associated with the choice of OS. Specifically, we believe that a partitioned operating system, when coupled with an appropriate architecture, can provide a strong infrastructure for developing systems for which reusability, modifiability, testability, and reliability are essential qualities. By adopting a partitioned OS, projects can gain benefits throughout the entire development lifecycle, from requirements and design, all the way to implementation, testing, and operations.

  19. Quality Assurance in Software Development: An Exploratory Investigation in Software Project Failures and Business Performance

    Science.gov (United States)

    Ichu, Emmanuel A.

    2010-01-01

    Software quality is perhaps one of the most sought-after attributes in product development, however; this goal is unattained. Problem factors in software development and how these have affected the maintainability of the delivered software systems requires a thorough investigation. It was, therefore, very important to understand software…

  20. ACEMAN (II): a PDP-11 software package for acoustic emission analysis

    International Nuclear Information System (INIS)

    Tobias, A.

    1976-01-01

    A powerful, but easy-to-use, software package (ACEMAN) for acoustic emission analysis has been developed at Berkeley Nuclear Laboratories. The system is based on a PDP-11 minicomputer with 24 K of memory, an RK05 DISK Drive and a Tektronix 4010 Graphics terminal. The operation of the system is described in detail in terms of the functions performed in response to the various command mnemonics. The ACEMAN software package offers many useful facilities not found on other acoustic emission monitoring systems. Its main features, many of which are unique, are summarised. The ACEMAN system automatically handles arrays of up to 12 sensors in real-time operation during which data are acquired, analysed, stored on the computer disk for future analysis and displayed on the terminal if required. (author)

  1. Developing Software Simulations

    Directory of Open Access Journals (Sweden)

    Tom Hall

    2007-06-01

    Full Text Available Programs in education and business often require learners to develop and demonstrate competence in specified areas and then be able to effectively apply this knowledge. One method to aid in developing a skill set in these areas is through the use of software simulations. These simulations can be used for learner demonstrations of competencies in a specified course as well as a review of the basic skills at the beginning of subsequent courses. The first section of this paper discusses ToolBook, the software used to develop our software simulations. The second section discusses the process of developing software simulations. The third part discusses how we have used software simulations to assess student knowledge of research design by providing simulations that allow the student to practice using SPSS and Excel.

  2. 48 CFR 227.7203-14 - Conformity, acceptance, and warranty of computer software and computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 3 2010-10-01 2010-10-01 false Conformity, acceptance... Software Documentation 227.7203-14 Conformity, acceptance, and warranty of computer software and computer...) Conformity and acceptance. Solicitations and contracts requiring the delivery of computer software shall...

  3. First-year experience with the ATLAS online monitoring framework

    International Nuclear Information System (INIS)

    Corso-Radu, A

    2010-01-01

    ATLAS is one of the four experiments in the Large Hadron Collider (LHC) at CERN, which has been put in operation this year. The challenging experimental environment and the extreme detector complexity required development of a highly scalable distributed monitoring framework, which is currently being used to monitor the quality of the data being taken as well as operational conditions of the hardware and software elements of the detector, trigger and data acquisition systems. At the moment the ATLAS Trigger/DAQ system is distributed over more than 1000 computers, which is about one third of the final ATLAS size. At every minute of an ATLAS data taking session the monitoring framework serves several thousands physics events to monitoring data analysis applications, handles more than 4 million histograms updates coming from more than 4 thousands applications, executes 10 thousands advanced data quality checks for a subset of those histograms, displays histograms and results of these checks on several dozens of monitors installed in main and satellite ATLAS control rooms. This note presents the overview of the online monitoring software framework, and describes the experience, which was gained during an extensive commissioning period as well as at the first phase of LHC beam in September 2008. Performance results, obtained on the current ATLAS DAQ system will also be presented, showing that the performance of the framework is adequate for the final ATLAS system.

  4. Input data requirements for performance modelling and monitoring of photovoltaic plants

    DEFF Research Database (Denmark)

    Gavriluta, Anamaria Florina; Spataru, Sergiu; Sera, Dezso

    2018-01-01

    This work investigates the input data requirements in the context of performance modeling of thin-film photovoltaic (PV) systems. The analysis focuses on the PVWatts performance model, well suited for on-line performance monitoring of PV strings, due to its low number of parameters and high......, modelling the performance of the PV modules at high irradiances requires a dataset of only a few hundred samples in order to obtain a power estimation accuracy of ~1-2\\%....

  5. Monitoring Assumptions in Assume-Guarantee Contracts

    Directory of Open Access Journals (Sweden)

    Oleg Sokolsky

    2016-05-01

    Full Text Available Pre-deployment verification of software components with respect to behavioral specifications in the assume-guarantee form does not, in general, guarantee absence of errors at run time. This is because assumptions about the environment cannot be discharged until the environment is fixed. An intuitive approach is to complement pre-deployment verification of guarantees, up to the assumptions, with post-deployment monitoring of environment behavior to check that the assumptions are satisfied at run time. Such a monitor is typically implemented by instrumenting the application code of the component. An additional challenge for the monitoring step is that environment behaviors are typically obtained through an I/O library, which may alter the component's view of the input format. This transformation requires us to introduce a second pre-deployment verification step to ensure that alarms raised by the monitor would indeed correspond to violations of the environment assumptions. In this paper, we describe an approach for constructing monitors and verifying them against the component assumption. We also discuss limitations of instrumentation-based monitoring and potential ways to overcome it.

  6. SOFTWARE OF MONITORING SYSTEM FOR ALLOCATED INFORMATION OF STATE PROGRAM ON INNOVATIVE DEVELOPMENT OF THE REPUBLIC OF BELARUS

    Directory of Open Access Journals (Sweden)

    V. A. Rybak

    2009-01-01

    Full Text Available The paper analyzes main indices (indicators of realization of the State program on innovative development of the Republic of Belarus (SPIDRB, contains and justifies a hierarchical structure of data processing and display, finalizes a list of the SPIDRB participants and executors. Major functions of the units pertaining to automation of an automated SPIDRB monitoring system are determined in the paper. In order to accumulate, process and furnish information a system of documentary databases on the basis of IBM Lotus Domino/Notes 8, relational databases of IBM DB2 and MS SQL Server 2005 is used in the given paper. Interaction with data suppliers is ensured by means of e-mail. The proposed scientific principles and software allow to automate a process of SPIDRB monitoring and to raise a decision-making efficiency in the field of innovative economic development of our country.

  7. Characterizing the contribution of quality requirements to software sustainability

    NARCIS (Netherlands)

    Condori-Fernandez, Nelly; Lago, Patricia

    2018-01-01

    Most respondents considered modifiability as relevant for addressing both technical and environmental sustainability. Functional correctness, availability, modifiability, interoperability and recoverability favor positively the endurability of software systems. This study has also identified

  8. Application Software for the Cabinet Operator Module of the Reactor Protection System

    International Nuclear Information System (INIS)

    Lee, Hyun-Chul; Jung, Hae-Won; Lee, Sung-Jin; Koo, Young-Ho; Kim, Seong-Tae; Kwak, Tae-Kil; Jin, Kyo-Hong

    2006-01-01

    A reactor protection system (RPS) plays the roles of generating the reactor trip signal and the engineered safety features (ESF) actuation signal when the monitored plant processes reach predefined limits. A Korean project group, so-called KNICS (Korean Nuclear I and C System), is developing a new digitalized RPS and the Cabinet Operator Module (COM) of the RPS which is used for the RPS integrity testing and monitoring by equipment operators. A flat panel display (FPD) with a touch screen capability is provided as a main user interface for the RPS. This paper shows the application software developed for the COM FPD. Equipment operators can monitor the status of the RPS and carry out various tests to verify system functions by means of the application software. A qualified hardware and software development environment are used to develop the application software

  9. A Design and Development of Multi-Purpose CCD Camera System with Thermoelectric Cooling: Software

    Directory of Open Access Journals (Sweden)

    S. H. Oh

    2007-12-01

    Full Text Available We present a software which we developed for the multi-purpose CCD camera. This software can be used on the all 3 types of CCD - KAF-0401E (768×512, KAF-1602E (15367times;1024, KAF-3200E (2184×1472 made in KODAK Co.. For the efficient CCD camera control, the software is operated with two independent processes of the CCD control program and the temperature/shutter operation program. This software is designed to fully automatic operation as well as manually operation under LINUX system, and is controled by LINUX user signal procedure. We plan to use this software for all sky survey system and also night sky monitoring and sky observation. As our results, the read-out time of each CCD are about 15sec, 64sec, 134sec for KAF-0401E, KAF-1602E, KAF-3200E., because these time are limited by the data transmission speed of parallel port. For larger format CCD, the data transmission is required more high speed. we are considering this control software to one using USB port for high speed data transmission.

  10. Engineering Runtime Requirements-Monitoring Systems Using MDA Technologies

    Science.gov (United States)

    Skene, James; Emmerich, Wolfgang

    The Model-Driven Architecture (MDA) technology toolset includes a language for describing the structure of meta-data, the MOF, and a language for describing consistency properties that data must exhibit, the OCL. Off-the-shelf tools can generate meta-data repositories and perform consistency checking over the data they contain. In this paper we describe how these tools can be used to implement runtime requirements monitoring of systems by modelling the required behaviour of the system, implementing a meta-data repository to collect system data, and consistency checking the repository to discover violations. We evaluate the approach by implementing a contract checker for the SLAng service-level agreement language, a language defined using a MOF meta-model, and integrating the checker into an Enterprise JavaBeans application. We discuss scalability issues resulting from immaturities in the applied technologies, leading to recommendations for their future development.

  11. Comparison of the Safety Critical Software V and V Requirements for the Research Reactor Instrumentation and Control System

    Energy Technology Data Exchange (ETDEWEB)

    Joo, Sungmoon; Suh, Yong-Suk; Park, Cheol [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2016-10-15

    This study was motivated by a research reactor project where the owner of the project and the equipment vendors are from two different standards frameworks. This paper reviews two major standards frameworks - NRC-IEEE and IAEA-IEC - and the software classification schemes as a background, then discuss the V and V issue. The purpose of this paper is by no means to solve the cross-standards-framework qualification issue, but, rather, is to remind the stakeholders of research reactor projects. V and V are also essential for the approval from regulatory bodies. As standards define or recommend consolidated engineering practices, methods, or criteria, V and V activities for software qualification are not exceptional. Within a standards framework, usually, the processes for the qualification of safety-critical software are well-established such that the safety is maximized while minimizing the compromises in software quality, safety, and reliability. When, however, multiple standards frameworks are involved in a research reactor project, it is difficult for equipment vendors to implement appropriate V and V activities as there is no unified view on this cross-standards-framework qualification issue yet. There are two major standards frameworks for safety-critical software development in nuclear industry. Unfortunately different safety classifications for software and thus different requirements for qualification are in place. What makes things worse is that (i) there are ambiguities in the standards and rooms for each stakeholders’ interpretation, and (ii) there is no one-to-one mapping between the associated V and V methods and activities. These may put the stakeholders of research reactor projects in trouble.

  12. Comparison of the Safety Critical Software V and V Requirements for the Research Reactor Instrumentation and Control System

    International Nuclear Information System (INIS)

    Joo, Sungmoon; Suh, Yong-Suk; Park, Cheol

    2016-01-01

    This study was motivated by a research reactor project where the owner of the project and the equipment vendors are from two different standards frameworks. This paper reviews two major standards frameworks - NRC-IEEE and IAEA-IEC - and the software classification schemes as a background, then discuss the V and V issue. The purpose of this paper is by no means to solve the cross-standards-framework qualification issue, but, rather, is to remind the stakeholders of research reactor projects. V and V are also essential for the approval from regulatory bodies. As standards define or recommend consolidated engineering practices, methods, or criteria, V and V activities for software qualification are not exceptional. Within a standards framework, usually, the processes for the qualification of safety-critical software are well-established such that the safety is maximized while minimizing the compromises in software quality, safety, and reliability. When, however, multiple standards frameworks are involved in a research reactor project, it is difficult for equipment vendors to implement appropriate V and V activities as there is no unified view on this cross-standards-framework qualification issue yet. There are two major standards frameworks for safety-critical software development in nuclear industry. Unfortunately different safety classifications for software and thus different requirements for qualification are in place. What makes things worse is that (i) there are ambiguities in the standards and rooms for each stakeholders’ interpretation, and (ii) there is no one-to-one mapping between the associated V and V methods and activities. These may put the stakeholders of research reactor projects in trouble

  13. Software essentials design and construction

    CERN Document Server

    Dingle, Adair

    2014-01-01

    About the Cover: Although capacity may be a problem for a doghouse, other requirements are usually minimal. Unlike skyscrapers, doghouses are simple units. They do not require plumbing, electricity, fire alarms, elevators, or ventilation systems, and they do not need to be built to code or pass inspections. The range of complexity in software design is similar. Given available software tools and libraries-many of which are free-hobbyists can build small or short-lived computer apps. Yet, design for software longevity, security, and efficiency can be intricate-as is the design of large-scale sy

  14. GIMS-Software for asset market experiments.

    Science.gov (United States)

    Palan, Stefan

    2015-03-01

    In this article we lay out requirements for an experimental market software for financial and economic research. We then discuss existing solutions. Finally, we introduce GIMS, an open source market software which is characterized by extensibility and ease of use, while offering nearly all of the required functionality.

  15. Standards Interoperability: Application of Contemporary Software Safety Assurance Standards to the Evolution of Legacy Software

    National Research Council Canada - National Science Library

    Meacham, Desmond J

    2006-01-01

    .... The proposed formal model is then applied to the requirements for RTCA DO-178B and MIL-STD-498 as representative examples of contemporary and legacy software standards. The results provide guidance on how to achieve airworthiness certification for modified legacy software, whilst maximizing the use of software products from the previous development.

  16. Supervised Semi-Automated Data Analysis Software for Gas Chromatography / Differential Mobility Spectrometry (GC/DMS) Metabolomics Applications.

    Science.gov (United States)

    Peirano, Daniel J; Pasamontes, Alberto; Davis, Cristina E

    2016-09-01

    Modern differential mobility spectrometers (DMS) produce complex and multi-dimensional data streams that allow for near-real-time or post-hoc chemical detection for a variety of applications. An active area of interest for this technology is metabolite monitoring for biological applications, and these data sets regularly have unique technical and data analysis end user requirements. While there are initial publications on how investigators have individually processed and analyzed their DMS metabolomic data, there are no user-ready commercial or open source software packages that are easily used for this purpose. We have created custom software uniquely suited to analyze gas chromatograph / differential mobility spectrometry (GC/DMS) data from biological sources. Here we explain the implementation of the software, describe the user features that are available, and provide an example of how this software functions using a previously-published data set. The software is compatible with many commercial or home-made DMS systems. Because the software is versatile, it can also potentially be used for other similarly structured data sets, such as GC/GC and other IMS modalities.

  17. Mobile Messaging Services-Based Personal Electrocardiogram Monitoring System

    Directory of Open Access Journals (Sweden)

    Ashraf A. Tahat

    2009-01-01

    Full Text Available A mobile monitoring system utilizing Bluetooth and mobile messaging services (MMS/SMSs with low-cost hardware equipment is proposed. A proof of concept prototype has been developed and implemented to enable transmission of an Electrocardiogram (ECG signal and body temperature of a patient, which can be expanded to include other vital signs. Communication between a mobile smart-phone and the ECG and temperature acquisition apparatus is implemented using the popular personal area network standard specification Bluetooth. When utilizing MMS for transmission, the mobile phone plots the received ECG signal and displays the temperature using special application software running on the client mobile phone itself, where the plot can be captured and saved as an image before transmission. Alternatively, SMS can be selected as a transmission means, where in this scenario, dedicated application software is required at the receiving device. The experimental setup can be operated for monitoring from anywhere in the globe covered by a cellular network that offers data services.

  18. Mobile messaging services-based personal electrocardiogram monitoring system.

    Science.gov (United States)

    Tahat, Ashraf A

    2009-01-01

    A mobile monitoring system utilizing Bluetooth and mobile messaging services (MMS/SMSs) with low-cost hardware equipment is proposed. A proof of concept prototype has been developed and implemented to enable transmission of an Electrocardiogram (ECG) signal and body temperature of a patient, which can be expanded to include other vital signs. Communication between a mobile smart-phone and the ECG and temperature acquisition apparatus is implemented using the popular personal area network standard specification Bluetooth. When utilizing MMS for transmission, the mobile phone plots the received ECG signal and displays the temperature using special application software running on the client mobile phone itself, where the plot can be captured and saved as an image before transmission. Alternatively, SMS can be selected as a transmission means, where in this scenario, dedicated application software is required at the receiving device. The experimental setup can be operated for monitoring from anywhere in the globe covered by a cellular network that offers data services.

  19. 40 CFR 141.563 - What follow-up action is my system required to take based on continuous turbidity monitoring?

    Science.gov (United States)

    2010-07-01

    ... required to take based on continuous turbidity monitoring? 141.563 Section 141.563 Protection of... Individual Filter Turbidity Requirements § 141.563 What follow-up action is my system required to take based on continuous turbidity monitoring? Follow-up action is required according to the following tables...

  20. SWEBOS – The Software Engineering Body of Skills

    Directory of Open Access Journals (Sweden)

    Yvonne Sedelmaier

    2015-02-01

    Full Text Available The development of complex software systems requires a mixture of various technical and non-technical competencies. While some guidelines exist which technical knowledge is required to make a good software engineer, there is a lack of insight as to which non-technical or soft skills are necessary to master complex software projects. This paper proposes a body of skills (SWEBOS for soft-ware engineering. The collection of necessary skills is developed on the basis of a clear, data-driven research design. The resulting required soft skills for software engineering are described precisely and semantically rich in a three-level structure. This approach guarantees that skills are not just characterized in a broad and general manner, but rather they are specifically adapted to the domain of software engineering.