WorldWideScience

Sample records for legacy software systems

  1. Architecture-driven Migration of Legacy Systems to Cloud-enabled Software

    DEFF Research Database (Denmark)

    Ahmad, Aakash; Babar, Muhammad Ali

    2014-01-01

    of legacy systems to cloud computing. The framework leverages the software reengineering concepts that aim to recover the architecture from legacy source code. Then the framework exploits the software evolution concepts to support architecture-driven migration of legacy systems to cloud-based architectures....... The Legacy-to-Cloud Migration Horseshoe comprises of four processes: (i) architecture migration planning, (ii) architecture recovery and consistency, (iii) architecture transformation and (iv) architecture-based development of cloud-enabled software. We aim to discover, document and apply the migration...

  2. Traceability of Software Safety Requirements in Legacy Safety Critical Systems

    Science.gov (United States)

    Hill, Janice L.

    2007-01-01

    How can traceability of software safety requirements be created for legacy safety critical systems? Requirements in safety standards are imposed most times during contract negotiations. On the other hand, there are instances where safety standards are levied on legacy safety critical systems, some of which may be considered for reuse for new applications. Safety standards often specify that software development documentation include process-oriented and technical safety requirements, and also require that system and software safety analyses are performed supporting technical safety requirements implementation. So what can be done if the requisite documents for establishing and maintaining safety requirements traceability are not available?

  3. Software Safety Risk in Legacy Safety-Critical Computer Systems

    Science.gov (United States)

    Hill, Janice L.; Baggs, Rhoda

    2007-01-01

    Safety Standards contain technical and process-oriented safety requirements. Technical requirements are those such as "must work" and "must not work" functions in the system. Process-Oriented requirements are software engineering and safety management process requirements. Address the system perspective and some cover just software in the system > NASA-STD-8719.13B Software Safety Standard is the current standard of interest. NASA programs/projects will have their own set of safety requirements derived from the standard. Safety Cases: a) Documented demonstration that a system complies with the specified safety requirements. b) Evidence is gathered on the integrity of the system and put forward as an argued case. [Gardener (ed.)] c) Problems occur when trying to meet safety standards, and thus make retrospective safety cases, in legacy safety-critical computer systems.

  4. Organisational scenarios and legacy systems

    OpenAIRE

    Brooke, Carole; Ramage, Magnus

    2001-01-01

    A legacy system is made up of technical components and social factors (such as software, people, skills, business processes) which no longer meet the needs of the business environment. The study of legacy systems has tended to be biased towards a software engineering perspective and to concentrate on technical properties. This paper suggests that the evaluation of potential change options for legacy systems can only be carried out as part of an holistic organisational analysis. That is, the e...

  5. Developing a TTCN-3 Test Harness for Legacy Software

    DEFF Research Database (Denmark)

    Okika, Joseph C.; Ravn, Anders Peter; Siddalingaiah, Lokesh

    2006-01-01

    We describe a prototype test harness for an embedded system which is the control software for a modern marine diesel engine. The operations of such control software requires complete certification. We adopt Testing and Test Control Notation (TTCN-3) to define test cases for this purpose. The main...... challenge in developing the test harness is to interface a generic test driver to the legacy software and provide a suitable interface for test engineers. The main contribution of this paper is a demonstration of a suitable design for such a test harness. It includes: a TTCN-3 test driver in C++, the legacy...

  6. Why Replacing Legacy Systems Is So Hard in Global Software Development: An Information Infrastructure Perspective

    DEFF Research Database (Denmark)

    Matthiesen, Stina; Bjørn, Pernille

    2015-01-01

    We report on an ethnographic study of an outsourcing global software development (GSD) setup between a Danish IT company and an Indian IT vendor developing a system to replace a legacy system for social services administration in Denmark. Physical distance and GSD collaboration issues tend...... to be obvious explanations for why GSD tasks fail to reach completion; however, we account for the difficulties within the technical nature of software system task. We use the framework of information infrastructure to show how replacing a legacy system in governmental information infrastructures includes...... the work of tracing back to knowledge concerning law, technical specifications, as well as how information infrastructures have dynamically evolved over time. Not easily carried out in a GSD setup is the work around technical tasks that requires careful examination of mundane technical aspects, standards...

  7. The ATLAS Trigger Simulation with Legacy Software

    CERN Document Server

    Bernius, Catrin; The ATLAS collaboration

    2017-01-01

    Physics analyses at the LHC require accurate simulations of the detector response and the event selection processes, generally done with the most recent software releases. The trigger response simulation is crucial for determination of overall selection efficiencies and signal sensitivities and should be done with the same software release with which data were recorded. This requires potentially running with software dating many years back, the so-called legacy software. Therefore having a strategy for running legacy software in a modern environment becomes essential when data simulated for past years start to present a sizeable fraction of the total. The requirements and possibilities for such a simulation scheme within the ATLAS software framework were examined and a proof-of-concept simulation chain has been successfully implemented. One of the greatest challenges was the choice of a data format which promises long term compatibility with old and new software releases. Over the time periods envisaged, data...

  8. Standards Interoperability: Application of Contemporary Software Safety Assurance Standards to the Evolution of Legacy Software

    National Research Council Canada - National Science Library

    Meacham, Desmond J

    2006-01-01

    .... The proposed formal model is then applied to the requirements for RTCA DO-178B and MIL-STD-498 as representative examples of contemporary and legacy software standards. The results provide guidance on how to achieve airworthiness certification for modified legacy software, whilst maximizing the use of software products from the previous development.

  9. The ATLAS Trigger Simulation with Legacy Software

    CERN Document Server

    Bernius, Catrin; The ATLAS collaboration

    2017-01-01

    Physics analyses at the LHC which search for rare physics processes or measure Standard Model parameters with high precision require accurate simulations of the detector response and the event selection processes. The accurate simulation of the trigger response is crucial for determination of overall selection efficiencies and signal sensitivities. For the generation and the reconstruction of simulated event data, generally the most recent software releases are used to ensure the best agreement between simulated data and real data. For the simulation of the trigger selection process, however, the same software release with which real data were taken should be ideally used. This requires potentially running with software dating many years back, the so-called legacy software. Therefore having a strategy for running legacy software in a modern environment becomes essential when data simulated for past years start to present a sizeable fraction of the total. The requirements and possibilities for such a simulatio...

  10. A Heuristic for Improving Legacy Software Quality during Maintenance: An Empirical Case Study

    Science.gov (United States)

    Sale, Michael John

    2017-01-01

    Many organizations depend on the functionality of mission-critical legacy software and the continued maintenance of this software is vital. Legacy software is defined here as software that contains no testing suite, is often foreign to the developer performing the maintenance, lacks meaningful documentation, and over time, has become difficult to…

  11. Bringing Legacy Visualization Software to Modern Computing Devices via Application Streaming

    Science.gov (United States)

    Fisher, Ward

    2014-05-01

    Planning software compatibility across forthcoming generations of computing platforms is a problem commonly encountered in software engineering and development. While this problem can affect any class of software, data analysis and visualization programs are particularly vulnerable. This is due in part to their inherent dependency on specialized hardware and computing environments. A number of strategies and tools have been designed to aid software engineers with this task. While generally embraced by developers at 'traditional' software companies, these methodologies are often dismissed by the scientific software community as unwieldy, inefficient and unnecessary. As a result, many important and storied scientific software packages can struggle to adapt to a new computing environment; for example, one in which much work is carried out on sub-laptop devices (such as tablets and smartphones). Rewriting these packages for a new platform often requires significant investment in terms of development time and developer expertise. In many cases, porting older software to modern devices is neither practical nor possible. As a result, replacement software must be developed from scratch, wasting resources better spent on other projects. Enabled largely by the rapid rise and adoption of cloud computing platforms, 'Application Streaming' technologies allow legacy visualization and analysis software to be operated wholly from a client device (be it laptop, tablet or smartphone) while retaining full functionality and interactivity. It mitigates much of the developer effort required by other more traditional methods while simultaneously reducing the time it takes to bring the software to a new platform. This work will provide an overview of Application Streaming and how it compares against other technologies which allow scientific visualization software to be executed from a remote computer. We will discuss the functionality and limitations of existing application streaming

  12. Modernization Solution for Legacy Banking System Using an Open Architecture

    Directory of Open Access Journals (Sweden)

    Constantin Marian MATEI

    2012-01-01

    Full Text Available Banks are still using legacy systems as the core of their business is comprised within such sys-tems. Since the technology and client demands are changing rapidly, the banks have to adapt their systems in order to be competitive. The issue is to identify correctly what are the bank users preferences in terms of software reliability and how modern is the system For instance, there are users who enjoy working using the old screen format, and there are users who enjoy working with newer layouts, Web interfaces, and so on. We need also to know the constraints generated by the usage of legacy systems, and how these systems can be improved or replaced. The scope of the article is to present a solution of modernizing a legacy banking system by using a SOA approach. The research is based on the modernization of a legacy system developed in COBOL/400 under IBM iSeries. The modernization process uses a SOA ap-proach using JAVA technologies.

  13. A methodology based on openEHR archetypes and software agents for developing e-health applications reusing legacy systems.

    Science.gov (United States)

    Cardoso de Moraes, João Luís; de Souza, Wanderley Lopes; Pires, Luís Ferreira; do Prado, Antonio Francisco

    2016-10-01

    In Pervasive Healthcare, novel information and communication technologies are applied to support the provision of health services anywhere, at anytime and to anyone. Since health systems may offer their health records in different electronic formats, the openEHR Foundation prescribes the use of archetypes for describing clinical knowledge in order to achieve semantic interoperability between these systems. Software agents have been applied to simulate human skills in some healthcare procedures. This paper presents a methodology, based on the use of openEHR archetypes and agent technology, which aims to overcome the weaknesses typically found in legacy healthcare systems, thereby adding value to the systems. This methodology was applied in the design of an agent-based system, which was used in a realistic healthcare scenario in which a medical staff meeting to prepare a cardiac surgery has been supported. We conducted experiments with this system in a distributed environment composed by three cardiology clinics and a center of cardiac surgery, all located in the city of Marília (São Paulo, Brazil). We evaluated this system according to the Technology Acceptance Model. The case study confirmed the acceptance of our agent-based system by healthcare professionals and patients, who reacted positively with respect to the usefulness of this system in particular, and with respect to task delegation to software agents in general. The case study also showed that a software agent-based interface and a tools-based alternative must be provided to the end users, which should allow them to perform the tasks themselves or to delegate these tasks to other people. A Pervasive Healthcare model requires efficient and secure information exchange between healthcare providers. The proposed methodology allows designers to build communication systems for the message exchange among heterogeneous healthcare systems, and to shift from systems that rely on informal communication of actors to

  14. Integrating commercial and legacy systems with EPICS

    International Nuclear Information System (INIS)

    Hill, J.O.; Kasemir, K.U.

    1997-01-01

    The Experimental Physics and Industrial Control System (EPICS) is a software toolkit, developed by a worldwide collaboration, which significantly reduces the level of effort required to implement a new control system. Recent developments now also significantly reduce the level of effort required to integrate commercial, legacy and/or site-authored control systems with EPICS. This paper will illustrate with an example both the level and type of effort required to use EPICS with other control system components as well as the benefits that may arise

  15. Software exorcism a handbook for debugging and optimizing legacy code

    CERN Document Server

    Blunden, Bill

    2013-01-01

    Software Exorcism: A Handbook for Debugging and Optimizing Legacy Code takes an unflinching, no bulls and look at behavioral problems in the software engineering industry, shedding much-needed light on the social forces that make it difficult for programmers to do their job. Do you have a co-worker who perpetually writes bad code that you are forced to clean up? This is your book. While there are plenty of books on the market that cover debugging and short-term workarounds for bad code, Reverend Bill Blunden takes a revolutionary step beyond them by bringing our atten

  16. Legacy systems: managing evolution through integration in a distributed and object-oriented computing environment.

    Science.gov (United States)

    Lemaitre, D; Sauquet, D; Fofol, I; Tanguy, L; Jean, F C; Degoulet, P

    1995-01-01

    Legacy systems are crucial for organizations since they support key functionalities. But they become obsolete with aging and the apparition of new techniques. Managing their evolution is a key issue in software engineering. This paper presents a strategy that has been developed at Broussais University Hospital in Paris to make a legacy system devoted to the management of health care units evolve towards a new up-to-date software. A two-phase evolution pathway is described. The first phase consists in separating the interface from the data storage and application control and in using a communication channel between the individualized components. The second phase proposes to use an object-oriented DBMS in place of the homegrown system. An application example for the management of hypertensive patients is described.

  17. Migration Performance for Legacy Data Access

    Directory of Open Access Journals (Sweden)

    Kam Woods

    2008-12-01

    Full Text Available We present performance data relating to the use of migration in a system we are creating to provide web access to heterogeneous document collections in legacy formats. Our goal is to enable sustained access to collections such as these when faced with increasing obsolescence of the necessary supporting applications and operating systems. Our system allows searching and browsing of the original files within their original contexts utilizing binary images of the original media. The system uses static and dynamic file migration to enhance collection browsing, and emulation to support both the use of legacy programs to access data and long-term preservation of the migration software. While we provide an overview of the architectural issues in building such a system, the focus of this paper is an in-depth analysis of file migration using data gathered from testing our software on 1,885 CD-ROMs and DVDs. These media are among the thousands of collections of social and scientific data distributed by the United States Government Printing Office (GPO on legacy media (CD-ROM, DVD, floppy disk under the Federal Depository Library Program (FDLP over the past 20 years.

  18. Transforming Cobol Legacy Software to a Generic Imperative Model

    National Research Council Canada - National Science Library

    Moraes, DinaL

    1999-01-01

    .... This research develops a transformation system to convert COBOL code into a generic imperative model, recapturing the initial design and deciphering the requirements implemented by the legacy code...

  19. Product Engineering Class in the Software Safety Risk Taxonomy for Building Safety-Critical Systems

    Science.gov (United States)

    Hill, Janice; Victor, Daniel

    2008-01-01

    When software safety requirements are imposed on legacy safety-critical systems, retrospective safety cases need to be formulated as part of recertifying the systems for further use and risks must be documented and managed to give confidence for reusing the systems. The SEJ Software Development Risk Taxonomy [4] focuses on general software development issues. It does not, however, cover all the safety risks. The Software Safety Risk Taxonomy [8] was developed which provides a construct for eliciting and categorizing software safety risks in a straightforward manner. In this paper, we present extended work on the taxonomy for safety that incorporates the additional issues inherent in the development and maintenance of safety-critical systems with software. An instrument called a Software Safety Risk Taxonomy Based Questionnaire (TBQ) is generated containing questions addressing each safety attribute in the Software Safety Risk Taxonomy. Software safety risks are surfaced using the new TBQ and then analyzed. In this paper we give the definitions for the specialized Product Engineering Class within the Software Safety Risk Taxonomy. At the end of the paper, we present the tool known as the 'Legacy Systems Risk Database Tool' that is used to collect and analyze the data required to show traceability to a particular safety standard

  20. The Legacy of Space Shuttle Flight Software

    Science.gov (United States)

    Hickey, Christopher J.; Loveall, James B.; Orr, James K.; Klausman, Andrew L.

    2011-01-01

    The initial goals of the Space Shuttle Program required that the avionics and software systems blaze new trails in advancing avionics system technology. Many of the requirements placed on avionics and software were accomplished for the first time on this program. Examples include comprehensive digital fly-by-wire technology, use of a digital databus for flight critical functions, fail operational/fail safe requirements, complex automated redundancy management, and the use of a high-order software language for flight software development. In order to meet the operational and safety goals of the program, the Space Shuttle software had to be extremely high quality, reliable, robust, reconfigurable and maintainable. To achieve this, the software development team evolved a software process focused on continuous process improvement and defect elimination that consistently produced highly predictable and top quality results, providing software managers the confidence needed to sign each Certificate of Flight Readiness (COFR). This process, which has been appraised at Capability Maturity Model (CMM)/Capability Maturity Model Integration (CMMI) Level 5, has resulted in one of the lowest software defect rates in the industry. This paper will present an overview of the evolution of the Primary Avionics Software System (PASS) project and processes over thirty years, an argument for strong statistical control of software processes with examples, an overview of the success story for identifying and driving out errors before flight, a case study of the few significant software issues and how they were either identified before flight or slipped through the process onto a flight vehicle, and identification of the valuable lessons learned over the life of the project.

  1. Security of legacy process control systems : Moving towards secure process control systems

    NARCIS (Netherlands)

    Oosterink, M.

    2012-01-01

    This white paper describes solutions which organisations may use to improve the security of their legacy process control systems. When we refer to a legacy system, we generally refer to old methodologies, technologies, computer systems or applications which are still in use, despite the fact that

  2. Integration of Sensor and Actuator Networks and the SCADA System to Promote the Migration of the Legacy Flexible Manufacturing System towards the Industry 4.0 Concept

    Directory of Open Access Journals (Sweden)

    Antonio José Calderón Godoy

    2018-05-01

    Full Text Available Networks of sensors and actuators in automated manufacturing processes are implemented using industrial fieldbuses, where automation units and supervisory systems are also connected to exchange operational information. In the context of the incoming fourth industrial revolution, called Industry 4.0, the management of legacy facilities is a paramount issue to deal with. This paper presents a solution to enhance the connectivity of a legacy Flexible Manufacturing System, which constitutes the first step in the adoption of the Industry 4.0 concept. Such a system includes the fieldbus PROcess FIeld BUS (PROFIBUS around which sensors, actuators, and controllers are interconnected. In order to establish effective communication between the sensors and actuators network and a supervisory system, a hardware and software approach including Ethernet connectivity is implemented. This work is envisioned to contribute to the migration of legacy systems towards the challenging Industry 4.0 framework. The experimental results prove the proper operation of the FMS and the feasibility of the proposal.

  3. ITSY Handheld Software Radio

    National Research Council Canada - National Science Library

    Bose, Vanu

    2001-01-01

    .... A handheld software radio platform would enable the construction of devices that could inter-operate with multiple legacy systems, download new waveforms and be used to construct adhoc networks...

  4. Experiences integrating autonomous components and legacy systems into tsunami early warning systems

    Science.gov (United States)

    Reißland, S.; Herrnkind, S.; Guenther, M.; Babeyko, A.; Comoglu, M.; Hammitzsch, M.

    2012-04-01

    Fostered by and embedded in the general development of Information and Communication Technology (ICT) the evolution of Tsunami Early Warning Systems (TEWS) shows a significant development from seismic-centred to multi-sensor system architectures using additional sensors, e.g. sea level stations for the detection of tsunami waves and GPS stations for the detection of ground displacements. Furthermore, the design and implementation of a robust and scalable service infrastructure supporting the integration and utilisation of existing resources serving near real-time data not only includes sensors but also other components and systems offering services such as the delivery of feasible simulations used for forecasting in an imminent tsunami threat. In the context of the development of the German Indonesian Tsunami Early Warning System (GITEWS) and the project Distant Early Warning System (DEWS) a service platform for both sensor integration and warning dissemination has been newly developed and demonstrated. In particular, standards of the Open Geospatial Consortium (OGC) and the Organization for the Advancement of Structured Information Standards (OASIS) have been successfully incorporated. In the project Collaborative, Complex, and Critical Decision-Support in Evolving Crises (TRIDEC) new developments are used to extend the existing platform to realise a component-based technology framework for building distributed TEWS. This talk will describe experiences made in GITEWS, DEWS and TRIDEC while integrating legacy stand-alone systems and newly developed special-purpose software components into TEWS using different software adapters and communication strategies to make the systems work together in a corporate infrastructure. The talk will also cover task management and data conversion between the different systems. Practical approaches and software solutions for the integration of sensors, e.g. providing seismic and sea level data, and utilisation of special

  5. Space Shuttle Program Primary Avionics Software System (PASS) Success Legacy - Quality and Reliability Date

    Science.gov (United States)

    Orr, James K.; Peltier, Daryl

    2010-01-01

    Thsi slide presentation reviews the avionics software system on board the space shuttle, with particular emphasis on the quality and reliability. The Primary Avionics Software System (PASS) provides automatic and fly-by-wire control of critical shuttle systems which executes in redundant computers. Charts given show the number of space shuttle flights vs time, PASS's development history, and other charts that point to the reliability of the system's development. The reliability of the system is also compared to predicted reliability.

  6. On the ergodic capacity of legacy systems in the presence of next generation interference

    KAUST Repository

    Mahmood, Nurul Huda

    2011-11-01

    Next generation wireless systems facilitating better utilization of the scarce radio spectrum have emerged as a response to inefficient rigid spectrum assignment policies. These are comprised of intelligent radio nodes that opportunistically operate in the radio spectrum of existing legacy systems; yet unwanted interference at the legacy receivers is unavoidable. In order to design efficient next generation systems and to minimize their harmful consequences, it is necessary to realize their impact on the performance of legacy systems. In this work, a generalized framework for the ergodic capacity analysis of such legacy systems in the presence of interference from next generation systems is presented. The analysis is built around a model developed for the statistical representation of the interference at the legacy receivers, which is then used to evaluate the ergodic capacity of the legacy system. Moreover, this analysis is not limited to the context of legacy systems, and is in fact applicaple to any interference limited system. Findings of analytical performance analyses are confirmed through selected computer-based Monte-Carlo simulations. © 2011 IEEE.

  7. Comparison of the Infiniti vision and the series 20,000 Legacy systems.

    Science.gov (United States)

    Fernández de Castro, Luis E; Solomon, Kerry D; Hu, Daniel J; Vroman, David T; Sandoval, Helga P

    2008-01-01

    To compare the efficiency of the Infiniti vision system and the Series 20,000 Legacy system phacoemulsification units during routine cataract extraction. Thirty-nine eyes of 39 patients were randomized to have their cataract removed using either the Infiniti or the Legacy system, both using the Neosonix handpiece. System settings were standardized. Ultrasound time, amount of balanced salt solution (BSS) used intraoperatively, and postoperative visual acuity at postoperative days 1, 7 and 30 were evaluated. Preoperatively, best corrected visual acuity was significantly worse in the Infiniti group compared to the Legacy group (0.38 +/- 0.23 and 0.21 +/- 0.16, respectively; p = 0.012). The mean phacoemulsification time was 39.6 +/- 22.9 s (range 6.0-102.0) for the Legacy group and 18.3 +/-19.1 s (range 1.0-80.0) for the Infiniti group (p = 0.001). The mean amounts of intraoperative BSS used were 117 +/- 37.7 ml (range 70-195) in the Legacy group and 85.3 +/- 38.9 ml (range 40-200) in the Infiniti group (p = 0.005). No differences in postoperative visual acuity were found. The ability to use higher flow rates and vacuum settings with the Infiniti vision system allowed for cataract removal with less phacoemulsification time than when using the Legacy system. Copyright 2008 S. Karger AG, Basel.

  8. Source Code Vulnerabilities in IoT Software Systems

    Directory of Open Access Journals (Sweden)

    Saleh Mohamed Alnaeli

    2017-08-01

    Full Text Available An empirical study that examines the usage of known vulnerable statements in software systems developed in C/C++ and used for IoT is presented. The study is conducted on 18 open source systems comprised of millions of lines of code and containing thousands of files. Static analysis methods are applied to each system to determine the number of unsafe commands (e.g., strcpy, strcmp, and strlen that are well-known among research communities to cause potential risks and security concerns, thereby decreasing a system’s robustness and quality. These unsafe statements are banned by many companies (e.g., Microsoft. The use of these commands should be avoided from the start when writing code and should be removed from legacy code over time as recommended by new C/C++ language standards. Each system is analyzed and the distribution of the known unsafe commands is presented. Historical trends in the usage of the unsafe commands of 7 of the systems are presented to show how the studied systems evolved over time with respect to the vulnerable code. The results show that the most prevalent unsafe command used for most systems is memcpy, followed by strlen. These results can be used to help train software developers on secure coding practices so that they can write higher quality software systems.

  9. A Transparent Translation from Legacy System Model into Common Information Model: Preprint

    Energy Technology Data Exchange (ETDEWEB)

    Ding, Fei [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Simpson, Jeffrey [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Zhang, Yingchen [National Renewable Energy Laboratory (NREL), Golden, CO (United States)

    2018-04-27

    Advance in smart grid is forcing utilities towards better monitoring, control and analysis of distribution systems, and requires extensive cyber-based intelligent systems and applications to realize various functionalities. The ability of systems, or components within systems, to interact and exchange services or information with each other is the key to the success of smart grid technologies, and it requires efficient information exchanging and data sharing infrastructure. The Common Information Model (CIM) is a standard that allows different applications to exchange information about an electrical system, and it has become a widely accepted solution for information exchange among different platforms and applications. However, most existing legacy systems are not developed using CIM, but using their own languages. Integrating such legacy systems is a challenge for utilities, and the appropriate utilization of the integrated legacy systems is even more intricate. Thus, this paper has developed an approach and open-source tool in order to translate legacy system models into CIM format. The developed tool is tested for a commercial distribution management system and simulation results have proved its effectiveness.

  10. Six Sigma software development

    CERN Document Server

    Tayntor, Christine B

    2002-01-01

    Since Six Sigma has had marked success in improving quality in other settings, and since the quality of software remains poor, it seems a natural evolution to apply the concepts and tools of Six Sigma to system development and the IT department. Until now however, there were no books available that applied these concepts to the system development process. Six Sigma Software Development fills this void and illustrates how Six Sigma concepts can be applied to all aspects of the evolving system development process. It includes the traditional waterfall model and in the support of legacy systems,

  11. 76 FR 15311 - Legacy Learning Systems, Inc.; Analysis of Proposed Consent Order To Aid Public Comment

    Science.gov (United States)

    2011-03-21

    ... FEDERAL TRADE COMMISSION [File No. 102 3055] Legacy Learning Systems, Inc.; Analysis of Proposed... electronically or in paper form. Comments should refer to ``Legacy Learning Systems, File No. 102 3055'' to... it. A comment filed in paper form should include the ``Legacy Learning Systems, File No. 102 3055...

  12. SQuAVisiT : A Software Quality Assessment and Visualisation Toolset

    NARCIS (Netherlands)

    Roubtsov, Serguei; Telea, Alexandru; Holten, Danny

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the Software Quality Assessment and Visualisation Toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully

  13. SQuAVisiT: a software quality assessment and visualisation toolset

    NARCIS (Netherlands)

    Roubtsov, S.; Telea, A.C.; Holten, D.H.R.

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the software quality assessment and visualisation toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully

  14. Getting Objects Methods and Interactions by Extracting Business Rules from Legacy Systems

    Directory of Open Access Journals (Sweden)

    Omar El Beggar

    2014-08-01

    Full Text Available The maintenance of legacy systems becomes over the years extremely complex and highly expensive due to the incessant changes of company activities and policies. In this case, a new or an improved system must replace the previous one. However, replacing those systems completely from scratch is also very expensive and it represents a huge risk. The optimal scenario is evolving those systems by profiting from the valuable knowledge embedded in them. This paper aims to present an approach for knowledge acquisition from existing legacy systems by extracting business rules from source code. In fact, the business rules are extracted and assigned next to the domain entities in order to generate objects methods and interactions in an object-oriented platform. Furthermore, a rules translation in natural language is given. The aim is advancing a solution for re-engineering legacy systems, minimize the cost of their modernization and keep very small the gap between the company business and the renovated systems.

  15. Domain management OSSs: bridging the gap between legacy and standards-based network management systems

    Science.gov (United States)

    Lemley, Todd A.

    1996-11-01

    The rapid change in the telecommunications environment is forcing carriers to re-assess not only their service offering, but also their network management philosophy. The competitive carrier environment has taken away the luxury of throwing technology at a problem by using legacy and proprietary systems and architectures. A more flexible management environment is necessary to effectively gain, and maintain operating margins in the new market era. Competitive forces are driving change which gives carriers more choices than those that are available in legacy and standards-based solutions alone. However, creating an operational support system (OSS) with this gap between legacy and standards has become as dynamic as the services which it supports. A philosophy which helps to integrate the legacy and standards systems is domain management. Domain management relates to a specific service or market 'domain,'and its associated operational support requirements. It supports a companies definition of its business model, which drives the definition of each domain. It also attempts to maximize current investment while injecting new technology available in a practical approach. The following paragraphs offer an overview of legacy systems, standards-based philosophy, and the potential of domain management to help bridge the gap between the two types of systems.

  16. An Architectural Framework for Integrating COTS/GOTS/Legacy Systems

    National Research Council Canada - National Science Library

    Gee, Karen

    2000-01-01

    .... To fully realize the DoD's goal, a new architectural framework is needed. This thesis proposes an architectural framework suitable for integrating COTS/GOTS/legacy systems in a distributed, heterogeneous environment...

  17. Monitoring extensions for component-based distributed software

    NARCIS (Netherlands)

    Diakov, N.K.; Papir, Z.; van Sinderen, Marten J.; Quartel, Dick

    2000-01-01

    This paper defines a generic class of monitoring extensions to component-based distributed enterprise software. Introducing a monitoring extension to a legacy application system can be very costly. In this paper, we identify the minimum support for application monitoring within the generic

  18. Legacy model integration for enhancing hydrologic interdisciplinary research

    Science.gov (United States)

    Dozier, A.; Arabi, M.; David, O.

    2013-12-01

    Many challenges are introduced to interdisciplinary research in and around the hydrologic science community due to advances in computing technology and modeling capabilities in different programming languages, across different platforms and frameworks by researchers in a variety of fields with a variety of experience in computer programming. Many new hydrologic models as well as optimization, parameter estimation, and uncertainty characterization techniques are developed in scripting languages such as Matlab, R, Python, or in newer languages such as Java and the .Net languages, whereas many legacy models have been written in FORTRAN and C, which complicates inter-model communication for two-way feedbacks. However, most hydrologic researchers and industry personnel have little knowledge of the computing technologies that are available to address the model integration process. Therefore, the goal of this study is to address these new challenges by utilizing a novel approach based on a publish-subscribe-type system to enhance modeling capabilities of legacy socio-economic, hydrologic, and ecologic software. Enhancements include massive parallelization of executions and access to legacy model variables at any point during the simulation process by another program without having to compile all the models together into an inseparable 'super-model'. Thus, this study provides two-way feedback mechanisms between multiple different process models that can be written in various programming languages and can run on different machines and operating systems. Additionally, a level of abstraction is given to the model integration process that allows researchers and other technical personnel to perform more detailed and interactive modeling, visualization, optimization, calibration, and uncertainty analysis without requiring deep understanding of inter-process communication. To be compatible, a program must be written in a programming language with bindings to a common

  19. Evolving software reengineering technology for the emerging innovative-competitive era

    Science.gov (United States)

    Hwang, Phillip Q.; Lock, Evan; Prywes, Noah

    1994-01-01

    This paper reports on a multi-tool commercial/military environment combining software Domain Analysis techniques with Reusable Software and Reengineering of Legacy Software. It is based on the development of a military version for the Department of Defense (DOD). The integrated tools in the military version are: Software Specification Assistant (SSA) and Software Reengineering Environment (SRE), developed by Computer Command and Control Company (CCCC) for Naval Surface Warfare Center (NSWC) and Joint Logistics Commanders (JLC), and the Advanced Research Project Agency (ARPA) STARS Software Engineering Environment (SEE) developed by Boeing for NAVAIR PMA 205. The paper describes transitioning these integrated tools to commercial use. There is a critical need for the transition for the following reasons: First, to date, 70 percent of programmers' time is applied to software maintenance. The work of these users has not been facilitated by existing tools. The addition of Software Reengineering will also facilitate software maintenance and upgrading. In fact, the integrated tools will support the entire software life cycle. Second, the integrated tools are essential to Business Process Reengineering, which seeks radical process innovations to achieve breakthrough results. Done well, process reengineering delivers extraordinary gains in process speed, productivity and profitability. Most importantly, it discovers new opportunities for products and services in collaboration with other organizations. Legacy computer software must be changed rapidly to support innovative business processes. The integrated tools will provide commercial organizations important competitive advantages. This, in turn, will increase employment by creating new business opportunities. Third, the integrated system will produce much higher quality software than use of the tools separately. The reason for this is that producing or upgrading software requires keen understanding of extremely complex

  20. Legacy Vehicle Fuel System Testing with Intermediate Ethanol Blends

    Energy Technology Data Exchange (ETDEWEB)

    Davis, G. W.; Hoff, C. J.; Borton, Z.; Ratcliff, M. A.

    2012-03-01

    The effects of E10 and E17 on legacy fuel system components from three common mid-1990s vintage vehicle models (Ford, GM, and Toyota) were studied. The fuel systems comprised a fuel sending unit with pump, a fuel rail and integrated pressure regulator, and the fuel injectors. The fuel system components were characterized and then installed and tested in sample aging test rigs to simulate the exposure and operation of the fuel system components in an operating vehicle. The fuel injectors were cycled with varying pulse widths during pump operation. Operational performance, such as fuel flow and pressure, was monitored during the aging tests. Both of the Toyota fuel pumps demonstrated some degradation in performance during testing. Six injectors were tested in each aging rig. The Ford and GM injectors showed little change over the aging tests. Overall, based on the results of both the fuel pump testing and the fuel injector testing, no major failures were observed that could be attributed to E17 exposure. The unknown fuel component histories add a large uncertainty to the aging tests. Acquiring fuel system components from operational legacy vehicles would reduce the uncertainty.

  1. Software ``Best'' Practices: Agile Deconstructed

    Science.gov (United States)

    Fraser, Steven

    Software “best” practices depend entirely on context - in terms of the problem domain, the system constructed, the software designers, and the “customers” ultimately deriving value from the system. Agile practices no longer have the luxury of “choosing” small non-mission critical projects with co-located teams. Project stakeholders are selecting and adapting practices based on a combina tion of interest, need and staffing. For example, growing product portfolios through a merger or the acquisition of a company exposes legacy systems to new staff, new software integration challenges, and new ideas. Innovation in communications (tools and processes) to span the growth and contraction of both information and organizations, while managing the adoption of changing software practices, is imperative for success. Traditional web-based tools such as web pages, document libraries, and forums are not suf ficient. A blend of tweeting, blogs, wikis, instant messaging, web-based confer encing, and telepresence creates a new dimension of communication “best” practices.

  2. Software for Collaborative Engineering of Launch Rockets

    Science.gov (United States)

    Stanley, Thomas Troy

    2003-01-01

    The Rocket Evaluation and Cost Integration for Propulsion and Engineering software enables collaborative computing with automated exchange of information in the design and analysis of launch rockets and other complex systems. RECIPE can interact with and incorporate a variety of programs, including legacy codes, that model aspects of a system from the perspectives of different technological disciplines (e.g., aerodynamics, structures, propulsion, trajectory, aeroheating, controls, and operations) and that are used by different engineers on different computers running different operating systems. RECIPE consists mainly of (1) ISCRM a file-transfer subprogram that makes it possible for legacy codes executed in their original operating systems on their original computers to exchange data and (2) CONES an easy-to-use filewrapper subprogram that enables the integration of legacy codes. RECIPE provides a tightly integrated conceptual framework that emphasizes connectivity among the programs used by the collaborators, linking these programs in a manner that provides some configuration control while facilitating collaborative engineering tradeoff studies, including design to cost studies. In comparison with prior collaborative-engineering schemes, one based on the use of RECIPE enables fewer engineers to do more in less time.

  3. Creation and implementation of the international information system for radiation legacy of the USSR 'RADLEG'

    International Nuclear Information System (INIS)

    Iskra, A.A.

    2002-01-01

    The stating of radiological problem of the radiation legacy of the Soviet and Russian military and civil programs of the nuclear fuel cycle have became possible after 'cold war' termination. The objective of the 'RADLEG' project is 'Development of a sophisticated computer based data system for evaluation of the radiation legacy of the former USSR and setting priorities on remediation and prevention policy'. The goal of the 'RADLEG' Project Phase 1 was creation of a simple operational database to be linked to GIS, describing currently available information on radiation legacy of the former USSR. During the Project Phase 2 the public accessible database linked to GIS has been developed. This GIS data system containing comprehensive information on the radiation legacy of the former Soviet Union has been developed in order to aid policy makers in two principle areas: to identify and set priorities on radiation safety problems, and to provide guidance for the development of technically, economically and socially sound policies to reduce health and environmental impact of radioactively contaminated sites. (author)

  4. Section 508 Electronic Information Accessibility Requirements for Software Development

    Science.gov (United States)

    Ellis, Rebecca

    2014-01-01

    Section 508 Subpart B 1194.21 outlines requirements for operating system and software development in order to create a product that is accessible to users with various disabilities. This portion of Section 508 contains a variety of standards to enable those using assistive technology and with visual, hearing, cognitive and motor difficulties to access all information provided in software. The focus on requirements was limited to the Microsoft Windows® operating system as it is the predominant operating system used at this center. Compliance with this portion of the requirements can be obtained by integrating the requirements into the software development cycle early and by remediating issues in legacy software if possible. There are certain circumstances with software that may arise necessitating an exemption from these requirements, such as design or engineering software using dynamically changing graphics or numbers to convey information. These exceptions can be discussed with the Section 508 Coordinator and another method of accommodation used.

  5. Long-term preservation of analysis software environment

    International Nuclear Information System (INIS)

    Toppe Larsen, Dag; Blomer, Jakob; Buncic, Predrag; Charalampidis, Ioannis; Haratyunyan, Artem

    2012-01-01

    Long-term preservation of scientific data represents a challenge to experiments, especially regarding the analysis software. Preserving data is not enough; the full software and hardware environment is needed. Virtual machines (VMs) make it possible to preserve hardware “in software”. A complete infrastructure package has been developed for easy deployment and management of VMs, based on CERN virtual machine (CernVM). Further, a HTTP-based file system, CernVM file system (CVMFS), is used for the distribution of the software. It is possible to process data with any given software version, and a matching, regenerated VM version. A point-and-click web user interface is being developed for setting up the complete processing chain, including VM and software versions, number and type of processing nodes, and the particular type of analysis and data. This paradigm also allows for distributed cloud-computing on private and public clouds, for both legacy and contemporary experiments.

  6. Legacy sample disposition project. Volume 2: Final report

    International Nuclear Information System (INIS)

    Gurley, R.N.; Shifty, K.L.

    1998-02-01

    This report describes the legacy sample disposition project at the Idaho Engineering and Environmental Laboratory (INEEL), which assessed Site-wide facilities/areas to locate legacy samples and owner organizations and then characterized and dispositioned these samples. This project resulted from an Idaho Department of Environmental Quality inspection of selected areas of the INEEL in January 1996, which identified some samples at the Test Reactor Area and Idaho Chemical Processing Plant that had not been characterized and dispositioned according to Resource Conservation and Recovery Act (RCRA) requirements. The objective of the project was to manage legacy samples in accordance with all applicable environmental and safety requirements. A systems engineering approach was used throughout the project, which included collecting the legacy sample information and developing a system for amending and retrieving the information. All legacy samples were dispositioned by the end of 1997. Closure of the legacy sample issue was achieved through these actions

  7. Resurrecting Legacy Code Using Ontosoft Knowledge-Sharing and Digital Object Management to Revitalize and Reproduce Software for Groundwater Management Research

    Science.gov (United States)

    Kwon, N.; Gentle, J.; Pierce, S. A.

    2015-12-01

    Software code developed for research is often used for a relatively short period of time before it is abandoned, lost, or becomes outdated. This unintentional abandonment of code is a valid problem in the 21st century scientific process, hindering widespread reusability and increasing the effort needed to develop research software. Potentially important assets, these legacy codes may be resurrected and documented digitally for long-term reuse, often with modest effort. Furthermore, the revived code may be openly accessible in a public repository for researchers to reuse or improve. For this study, the research team has begun to revive the codebase for Groundwater Decision Support System (GWDSS), originally developed for participatory decision making to aid urban planning and groundwater management, though it may serve multiple use cases beyond those originally envisioned. GWDSS was designed as a java-based wrapper with loosely federated commercial and open source components. If successfully revitalized, GWDSS will be useful for both practical applications as a teaching tool and case study for groundwater management, as well as informing theoretical research. Using the knowledge-sharing approaches documented by the NSF-funded Ontosoft project, digital documentation of GWDSS is underway, from conception to development, deployment, characterization, integration, composition, and dissemination through open source communities and geosciences modeling frameworks. Information assets, documentation, and examples are shared using open platforms for data sharing and assigned digital object identifiers. Two instances of GWDSS version 3.0 are being created: 1) a virtual machine instance for the original case study to serve as a live demonstration of the decision support tool, assuring the original version is usable, and 2) an open version of the codebase, executable installation files, and developer guide available via an open repository, assuring the source for the

  8. Software system safety

    Science.gov (United States)

    Uber, James G.

    1988-01-01

    Software itself is not hazardous, but since software and hardware share common interfaces there is an opportunity for software to create hazards. Further, these software systems are complex, and proven methods for the design, analysis, and measurement of software safety are not yet available. Some past software failures, future NASA software trends, software engineering methods, and tools and techniques for various software safety analyses are reviewed. Recommendations to NASA are made based on this review.

  9. The current system of higher education in India inherits the legacy

    Indian Academy of Sciences (India)

    The current system of higher education in India inherits the legacy of colonial proposals and legislations dating back to the early 19th century. The social sciences and humanities still carry disciplinary burdens that need revisiting as we attempt to think of new educational strategies for the 21st century.

  10. The Java Legacy Interface

    DEFF Research Database (Denmark)

    Korsholm, Stephan

    2007-01-01

    The Java Legacy Interface is designed to use Java for encapsulating native legacy code on small embedded platforms. We discuss why existing technologies for encapsulating legacy code (JNI) is not sufficient for an important range of small embedded platforms, and we show how the Java Legacy...... Interface offers this previously missing functionality. We describe an implementation of the Java Legacy Interface for a particular virtual machine, and how we have used this virtual machine to integrate Java with an existing, commercial, soft real-time, C/C++ legacy platform....

  11. Application Programmer's Interface (API) for Heterogeneous Language Environment and Upgrading the Legacy Embedded Software

    National Research Council Canada - National Science Library

    Moua, Theng

    2001-01-01

    .... The shortage of original software designs, lack of corporate knowledge and software design documentation, unsupported programming languages, and obsolete real-time operating system and development...

  12. Report of experiments and evidence for ASC L2 milestone 4467 : demonstration of a legacy application's path to exascale.

    Energy Technology Data Exchange (ETDEWEB)

    Curry, Matthew L.; Ferreira, Kurt Brian; Pedretti, Kevin Thomas Tauke; Leung, Vitus Joseph; Moreland, Kenneth D.; Lofstead, Gerald Fredrick, II; Gentile, Ann C. (Sandia National Laboratories, Livermore, CA); Klundt, Ruth Ann; Ward, H. Lee; Laros, James H., III; Hemmert, Karl Scott; Fabian, Nathan D.; Levenhagen, Michael J.; Barrett, Brian W.; Brightwell, Ronald Brian; Barrett, Richard; Wheeler, Kyle Bruce; Kelly, Suzanne Marie; Rodrigues, Arun F.; Brandt, James M. (Sandia National Laboratories, Livermore, CA); Thompson, David (Sandia National Laboratories, Livermore, CA); VanDyke, John P.; Oldfield, Ron A.; Tucker, Thomas (Open Grid Computing, Inc., Austin, TX); Vaughan, Courtenay Thomas

    2012-03-01

    This report documents thirteen of Sandia's contributions to the Computational Systems and Software Environment (CSSE) within the Advanced Simulation and Computing (ASC) program between fiscal years 2009 and 2012. It describes their impact on ASC applications. Most contributions are implemented in lower software levels allowing for application improvement without source code changes. Improvements are identified in such areas as reduced run time, characterizing power usage, and Input/Output (I/O). Other experiments are more forward looking, demonstrating potential bottlenecks using mini-application versions of the legacy codes and simulating their network activity on Exascale-class hardware. The purpose of this report is to prove that the team has completed milestone 4467-Demonstration of a Legacy Application's Path to Exascale. Cielo is expected to be the last capability system on which existing ASC codes can run without significant modifications. This assertion will be tested to determine where the breaking point is for an existing highly scalable application. The goal is to stretch the performance boundaries of the application by applying recent CSSE RD in areas such as resilience, power, I/O, visualization services, SMARTMAP, lightweight LWKs, virtualization, simulation, and feedback loops. Dedicated system time reservations and/or CCC allocations will be used to quantify the impact of system-level changes to extend the life and performance of the ASC code base. Finally, a simulation of anticipated exascale-class hardware will be performed using SST to supplement the calculations. Determine where the breaking point is for an existing highly scalable application: Chapter 15 presented the CSSE work that sought to identify the breaking point in two ASC legacy applications-Charon and CTH. Their mini-app versions were also employed to complete the task. There is no single breaking point as more than one issue was found with the two codes. The results were

  13. System support software for TSTA

    International Nuclear Information System (INIS)

    Claborn, G.W.; Mann, L.W.; Nielson, C.W.

    1987-01-01

    The software at the Tritium Systems Test Assembly (TSTA) is logically broken into two parts, the system support software and the subsystem software. The purpose of the system support software is to isolate the subsystem software from the physical hardware. In this sense the system support software forms the kernel of the software at TSTA. The kernel software performs several functions. It gathers data from CAMAC modules and makes that data available for subsystem processes. It services requests to send commands to CAMAC modules. It provides a system of logging functions and provides for a system-wide global program state that allows highly structured interaction between subsystem processes. The kernel's most visible function is to provide the Man-Machine Interface (MMI). The MMI allows the operators a window into the physical hardware and subsystem process state. Finally the kernel provides a data archiving and compression function that allows archival data to be accessed and plotted. Such kernel software as developed and implemented at TSTA is described

  14. Software quality assurance: in large scale and complex software-intensive systems

    NARCIS (Netherlands)

    Mistrik, I.; Soley, R.; Ali, N.; Grundy, J.; Tekinerdogan, B.

    2015-01-01

    Software Quality Assurance in Large Scale and Complex Software-intensive Systems presents novel and high-quality research related approaches that relate the quality of software architecture to system requirements, system architecture and enterprise-architecture, or software testing. Modern software

  15. Assessing Resistance to Change During Shifting from Legacy to Open Web-Based Systems in the Air Transport Industry

    Science.gov (United States)

    Brewer, Denise

    The air transport industry (ATI) is a dynamic, communal, international, and intercultural environment in which the daily operations of airlines, airports, and service providers are dependent on information technology (IT). Many of the IT legacy systems are more than 30 years old, and current regulations and the globally distributed workplace have brought profound changes to the way the ATI community interacts. The purpose of the study was to identify the areas of resistance to change in the ATI community and the corresponding factors in change management requirements that minimize product development delays and lead to a successful and timely shift from legacy to open web-based systems in upgrading ATI operations. The research questions centered on product development team processes as well as the members' perceived need for acceptance of change. A qualitative case study approach rooted in complexity theory was employed using a single case of an intercultural product development team dispersed globally. Qualitative data gathered from questionnaires were organized using Nvivo software, which coded the words and themes. Once coded, themes emerged identifying the areas of resistance within the product development team. Results of follow-up interviews with team members suggests that intercultural relationship building prior to and during project execution; focus on common team goals; and, development of relationships to enhance interpersonal respect, understanding and overall communication help overcome resistance to change. Positive social change in the form of intercultural group effectiveness evidenced in increased team functioning during major project transitions is likely to result when global managers devote time to cultural understanding.

  16. Pushouts in software architecture design

    OpenAIRE

    Riché, T. L.; Gonçalves, Rui; Marker, B.; Batory, D.

    2012-01-01

    A classical approach to program derivation is to progressively extend a simple specification and then incrementally refine it to an implementation. We claim this approach is hard or impractical when reverse engineering legacy software architectures. We present a case study that shows optimizations and pushouts--in addition to refinements and extensions--are essential for practical stepwise development of complex software architectures. NSF CCF 0724979 NSF CNS 0509338 NSF CCF 0917167 ...

  17. The architecture of a reliable software monitoring system for embedded software systems

    International Nuclear Information System (INIS)

    Munson, J.; Krings, A.; Hiromoto, R.

    2006-01-01

    We develop the notion of a measurement-based methodology for embedded software systems to ensure properties of reliability, survivability and security, not only under benign faults but under malicious and hazardous conditions as well. The driving force is the need to develop a dynamic run-time monitoring system for use in these embedded mission critical systems. These systems must run reliably, must be secure and they must fail gracefully. That is, they must continue operating in the face of the departures from their nominal operating scenarios, the failure of one or more system components due to normal hardware and software faults, as well as malicious acts. To insure the integrity of embedded software systems, the activity of these systems must be monitored as they operate. For each of these systems, it is possible to establish a very succinct representation of nominal system activity. Furthermore, it is possible to detect departures from the nominal operating scenario in a timely fashion. Such departure may be due to various circumstances, e.g., an assault from an outside agent, thus forcing the system to operate in an off-nominal environment for which it was neither tested nor certified, or a hardware/software component that has ceased to operate in a nominal fashion. A well-designed system will have the property of graceful degradation. It must continue to run even though some of the functionality may have been lost. This involves the intelligent re-mapping of system functions. Those functions that are impacted by the failure of a system component must be identified and isolated. Thus, a system must be designed so that its basic operations may be re-mapped onto system components still operational. That is, the mission objectives of the software must be reassessed in terms of the current operational capabilities of the software system. By integrating the mechanisms to support observation and detection directly into the design methodology, we propose to shift

  18. Hardware and software architecture for the integration of the new EC waves launcher in FTU control system

    Energy Technology Data Exchange (ETDEWEB)

    Boncagni, L. [Associazione EURATOM-ENEA sulla Fusione – ENEA, Via Enrico Fermi, 45 00045 Frascati (RM) (Italy); Centioli, C., E-mail: cristina.centioli@enea.it [Associazione EURATOM-ENEA sulla Fusione – ENEA, Via Enrico Fermi, 45 00045 Frascati (RM) (Italy); Galperti, C.; Alessi, E.; Granucci, G. [Associazione EURATOM-ENEA-CNR sulla Fusione – IFP-CNR, Via Roberto Cozzi, 53 20125 Milano (Italy); Grosso, L.A. [Associazione EURATOM-ENEA sulla Fusione – ENEA, Via Enrico Fermi, 45 00045 Frascati (RM) (Italy); Marchetto, C. [Associazione EURATOM-ENEA-CNR sulla Fusione – IFP-CNR, Via Roberto Cozzi, 53 20125 Milano (Italy); Napolitano, M. [Associazione EURATOM-ENEA sulla Fusione – ENEA, Via Enrico Fermi, 45 00045 Frascati (RM) (Italy); Nowak, S. [Associazione EURATOM-ENEA-CNR sulla Fusione – IFP-CNR, Via Roberto Cozzi, 53 20125 Milano (Italy); Panella, M. [Associazione EURATOM-ENEA sulla Fusione – ENEA, Via Enrico Fermi, 45 00045 Frascati (RM) (Italy); Sozzi, C. [Associazione EURATOM-ENEA-CNR sulla Fusione – IFP-CNR, Via Roberto Cozzi, 53 20125 Milano (Italy); Tilia, B.; Vitale, V. [Associazione EURATOM-ENEA sulla Fusione – ENEA, Via Enrico Fermi, 45 00045 Frascati (RM) (Italy)

    2013-10-15

    Highlights: ► The integration of a new ECRH launcher to FTU legacy control system is reported. ► Fast control has been developed with a three-node RT cluster within MARTe framework. ► Slow control was implemented with a Simatic S7 PLC and an EPICS IOC-CA application. ► The first results have assessed the feasibility of the launcher control architecture. -- Abstract: The role of high power electron cyclotron (EC) waves in controlling magnetohydrodynamic (MHD) instabilities in tokamaks has been assessed in several experiments, exploiting the physical effects induced by resonant heating and current drive. Recently a new EC launcher, whose main goal is controlling tearing modes and possibly preventing their onset, is being implemented on FTU. So far most of the components of the launcher control strategy have been realized and successfully tested on plasma experiments. Nevertheless the operations of the new launcher must be completely integrated into the existing one, and to FTU control system. This work deals with this final step, proposing a hardware and software architecture implementing up to date technologies, to achieve a modular and effective control strategy well integrated into a legacy system. The slow control system of the new EC launcher is based on a Siemens S7 Programmable Logic Controller (PLC), integrated into FTU control system supervisor through an EPICS input output controller (IOC) and an in-house developed Channel Access client application creating an abstraction layer that decouples the IOC and the PLC from the FTU Supervisor software. This architecture could enable a smooth migration to an EPICS-only supervisory control system. The real time component of the control system is based on the open source MARTe framework relying on a Linux real time cluster, devoted to the detection of MHD instabilities and the calculation of the injection angles and the time reference for the radiofrequency power enable commands for the EC launcher.

  19. Virtualization of Legacy Instrumentation Control Computers for Improved Reliability, Operational Life, and Management.

    Science.gov (United States)

    Katz, Jonathan E

    2017-01-01

    Laboratories tend to be amenable environments for long-term reliable operation of scientific measurement equipment. Indeed, it is not uncommon to find equipment 5, 10, or even 20+ years old still being routinely used in labs. Unfortunately, the Achilles heel for many of these devices is the control/data acquisition computer. Often these computers run older operating systems (e.g., Windows XP) and, while they might only use standard network, USB or serial ports, they require proprietary software to be installed. Even if the original installation disks can be found, it is a burdensome process to reinstall and is fraught with "gotchas" that can derail the process-lost license keys, incompatible hardware, forgotten configuration settings, etc. If you have running legacy instrumentation, the computer is the ticking time bomb waiting to put a halt to your operation.In this chapter, I describe how to virtualize your currently running control computer. This virtualized computer "image" is easy to maintain, easy to back up and easy to redeploy. I have used this multiple times in my own lab to greatly improve the robustness of my legacy devices.After completing the steps in this chapter, you will have your original control computer as well as a virtual instance of that computer with all the software installed ready to control your hardware should your original computer ever be decommissioned.

  20. The SIRTF Legacy Observing Program

    Science.gov (United States)

    Greenhouse, M. A.; Leisawitz, D.; Gehrz, R. D.; Clemens, D. P.; Force, Sirtf Community Task

    1997-12-01

    Legacy Observations and General Observations(GO)are separate categories in which SIRTF observing time will be allocated through peer reviewed community proposals. The Legacy Program will embrace several projects, each headed by a Legacy Principal Investigator. Legacy Observations are distinguished from General Observations by the following three criteria: [1] the project is a large, coherent investigation whose scientific goals can not be met by a number of smaller, uncoordinated projects; [2] the data will be of both general and lasting importance to the broad astronomical community and of immediate utility in motivating and planning follow-on GO investigations with SIRTF; and [3] the data (unprocessed, fully processed, and at intermediate steps in processing) will be placed in a public data base immediately and with no proprietary period. The goals of the SIRTF Legacy program are: [1] enable community use of SIRTF for large coherent survey observations, [2] provide prompt community access to SIRTF survey data, and [3] enable GO program observations based on Legacy program results. A likely attribute (but not a requirement) for Legacy projects is that they may involve hundreds, and perhaps thousands, of hours of observing time. It is anticipated that as much as 6000 hours of telescope time will be allocated through the Legacy program. To meet Legacy program goal [3], allocation of as much as 70% of SIRTF's first year on orbit to Legacy projects may be necessary, and the observing phase of the Legacy program will be completed during the following year. A Legacy call for proposals will be issued 1 year prior to launch or sooner, and will be open to all scientists and science topics. In this poster, we display Legacy program definition and schedule items that will be of interest to those intending to propose under this unique opportunity.

  1. The Planck Legacy Archive

    Science.gov (United States)

    Dupac, X.; Arviset, C.; Fernandez Barreiro, M.; Lopez-Caniego, M.; Tauber, J.

    2015-12-01

    The Planck Collaboration has released in 2015 their second major dataset through the Planck Legacy Archive (PLA). It includes cosmological, Extragalactic and Galactic science data in temperature (intensity) and polarization. Full-sky maps are provided with unprecedented angular resolution and sensitivity, together with a large number of ancillary maps, catalogues (generic, SZ clusters and Galactic cold clumps), time-ordered data and other information. The extensive cosmological likelihood package allows cosmologists to fully explore the plausible parameters of the Universe. A new web-based PLA user interface is made public since Dec. 2014, allowing easier and faster access to all Planck data, and replacing the previous Java-based software. Numerous additional improvements to the PLA are also being developed through the so-called PLA Added-Value Interface, making use of an external contract with the Planetek Hellas and Expert Analytics software companies. This will allow users to process time-ordered data into sky maps, separate astrophysical components in existing maps, simulate the microwave and infrared sky through the Planck Sky Model, and use a number of other functionalities.

  2. Secure Wireless Military Healthcare Telemedicine Enterprise System

    National Research Council Canada - National Science Library

    Lucas, Kenneth

    2003-01-01

    ...(exp TM) software and Division Tools with cross platform telemedicine systems, inclusive of computer based systems, handheld wireless PDA devices, and miniature computers, to existing DoD legacy...

  3. Secure Wireless Military Healthcare Telemedicine Enterprise System

    National Research Council Canada - National Science Library

    Lucas, Kenneth

    2002-01-01

    ...) software and Dvision Tools with cross platform telemedicine systems, inclusive of computer based systems, handheld wireless PDA devices, and miniature computers, to existing DoD legacy and developing...

  4. One-Click Data Analysis Software for Science Operations

    Science.gov (United States)

    Navarro, Vicente

    2015-12-01

    One of the important activities of ESA Science Operations Centre is to provide Data Analysis Software (DAS) to enable users and scientists to process data further to higher levels. During operations and post-operations, Data Analysis Software (DAS) is fully maintained and updated for new OS and library releases. Nonetheless, once a Mission goes into the "legacy" phase, there are very limited funds and long-term preservation becomes more and more difficult. Building on Virtual Machine (VM), Cloud computing and Software as a Service (SaaS) technologies, this project has aimed at providing long-term preservation of Data Analysis Software for the following missions: - PIA for ISO (1995) - SAS for XMM-Newton (1999) - Hipe for Herschel (2009) - EXIA for EXOSAT (1983) Following goals have guided the architecture: - Support for all operations, post-operations and archive/legacy phases. - Support for local (user's computer) and cloud environments (ESAC-Cloud, Amazon - AWS). - Support for expert users, requiring full capabilities. - Provision of a simple web-based interface. This talk describes the architecture, challenges, results and lessons learnt gathered in this project.

  5. 31 CFR 357.20 - Securities account in Legacy Treasury Direct ®.

    Science.gov (United States)

    2010-07-01

    ... 31 Money and Finance: Treasury 2 2010-07-01 2010-07-01 false Securities account in Legacy Treasury Direct ®. 357.20 Section 357.20 Money and Finance: Treasury Regulations Relating to Money and Finance... Securities System (Legacy Treasury Direct) § 357.20 Securities account in Legacy Treasury Direct ®. (a...

  6. Recommendation systems in software engineering

    CERN Document Server

    Robillard, Martin P; Walker, Robert J; Zimmermann, Thomas

    2014-01-01

    With the growth of public and private data stores and the emergence of off-the-shelf data-mining technology, recommendation systems have emerged that specifically address the unique challenges of navigating and interpreting software engineering data.This book collects, structures and formalizes knowledge on recommendation systems in software engineering. It adopts a pragmatic approach with an explicit focus on system design, implementation, and evaluation. The book is divided into three parts: "Part I - Techniques" introduces basics for building recommenders in software engineering, including techniques for collecting and processing software engineering data, but also for presenting recommendations to users as part of their workflow.?"Part II - Evaluation" summarizes methods and experimental designs for evaluating recommendations in software engineering.?"Part III - Applications" describes needs, issues and solution concepts involved in entire recommendation systems for specific software engineering tasks, fo...

  7. Identifying the Return on Investment for Army Migration to a Modular Open Systems Approach for Future and Legacy Systems

    Science.gov (United States)

    2017-04-05

    Identifying the Return on Investment for Army Migration to a Modular Open Systems Approach for Future and Legacy Systems Phillip Minor...Authorization Act (NDAA) of 2015, cites the modular open systems approach (MOSA) as both a business and technical strategy to reduce the cost of system ...access the service over the network. Combine the advances cited above with the emergence of systems developed using the modular open systems approach

  8. Security System Software

    Science.gov (United States)

    1993-01-01

    C Language Integration Production System (CLIPS), a NASA-developed expert systems program, has enabled a security systems manufacturer to design a new generation of hardware. C.CURESystem 1 Plus, manufactured by Software House, is a software based system that is used with a variety of access control hardware at installations around the world. Users can manage large amounts of information, solve unique security problems and control entry and time scheduling. CLIPS acts as an information management tool when accessed by C.CURESystem 1 Plus. It asks questions about the hardware and when given the answer, recommends possible quick solutions by non-expert persons.

  9. Common software for controlling and monitoring the upgraded CMS Level-1 trigger

    CERN Document Server

    Codispoti, Giuseppe

    2017-01-01

    The Large Hadron Collider restarted in 2015 with a higher centre-of-mass energy of 13 TeV. The instantaneous luminosity is expected to increase significantly in the coming years. An upgraded Level-1 trigger system was deployed in the CMS experiment in order to maintain the same efficiencies for searches and precision measurements as those achieved in 2012. This system must be controlled and monitored coherently through software, with high operational efficiency.The legacy system was composed of a large number of custom data processor boards; correspondingly, only a small fraction of the software was common between the different subsystems. The upgraded system is composed of a set of general purpose boards, that follow the MicroTCA specification, and transmit data over optical links, resulting in a more homogeneous system. The associated software is based on generic components corresponding to the firmware blocks that are shared across different cards, regardless of the role that the card plays in the system. ...

  10. An Analysis of Electronic Commerce Acquisition Systems: Comparison of a New Pure Electronic Purchasing and Exchange System (Electronic Storefront) and Other Legacy On-line Purchasing Systems

    National Research Council Canada - National Science Library

    Rowe, Arthur

    2002-01-01

    ... as they relate to contracting and purchasing of supplies and services, The issues and concerns with legacy on-line procurement systems will be compared to a newly developed Pure Electronic Ordering System...

  11. Software quality assurance and software safety in the Biomed Control System

    International Nuclear Information System (INIS)

    Singh, R.P.; Chu, W.T.; Ludewigt, B.A.; Marks, K.M.; Nyman, M.A.; Renner, T.R.; Stradtner, R.

    1989-01-01

    The Biomed Control System is a hardware/software system used for the delivery, measurement and monitoring of heavy-ion beams in the patient treatment and biology experiment rooms in the Bevalac at the Lawrence Berkeley Laboratory (LBL). This paper describes some aspects of this system including historical background philosophy, configuration management, hardware features that facilitate software testing, software testing procedures, the release of new software quality assurance, safety and operator monitoring. 3 refs

  12. Terra Harvest software architecture

    Science.gov (United States)

    Humeniuk, Dave; Klawon, Kevin

    2012-06-01

    Under the Terra Harvest Program, the DIA has the objective of developing a universal Controller for the Unattended Ground Sensor (UGS) community. The mission is to define, implement, and thoroughly document an open architecture that universally supports UGS missions, integrating disparate systems, peripherals, etc. The Controller's inherent interoperability with numerous systems enables the integration of both legacy and future UGS System (UGSS) components, while the design's open architecture supports rapid third-party development to ensure operational readiness. The successful accomplishment of these objectives by the program's Phase 3b contractors is demonstrated via integration of the companies' respective plug-'n'-play contributions that include controllers, various peripherals, such as sensors, cameras, etc., and their associated software drivers. In order to independently validate the Terra Harvest architecture, L-3 Nova Engineering, along with its partner, the University of Dayton Research Institute, is developing the Terra Harvest Open Source Environment (THOSE), a Java Virtual Machine (JVM) running on an embedded Linux Operating System. The Use Cases on which the software is developed support the full range of UGS operational scenarios such as remote sensor triggering, image capture, and data exfiltration. The Team is additionally developing an ARM microprocessor-based evaluation platform that is both energy-efficient and operationally flexible. The paper describes the overall THOSE architecture, as well as the design decisions for some of the key software components. Development process for THOSE is discussed as well.

  13. Multi-Level Formation of Complex Software Systems

    Directory of Open Access Journals (Sweden)

    Hui Li

    2016-05-01

    Full Text Available We present a multi-level formation model for complex software systems. The previous works extract the software systems to software networks for further studies, but usually investigate the software networks at the class level. In contrast to these works, our treatment of software systems as multi-level networks is more realistic. In particular, the software networks are organized by three levels of granularity, which represents the modularity and hierarchy in the formation process of real-world software systems. More importantly, simulations based on this model have generated more realistic structural properties of software networks, such as power-law, clustering and modularization. On the basis of this model, how the structure of software systems effects software design principles is then explored, and it could be helpful for understanding software evolution and software engineering practices.

  14. Software And Systems Engineering Risk Management

    Science.gov (United States)

    2010-04-01

    RSKM 2004 COSO Enterprise RSKM Framework 2006 ISO/IEC 16085 Risk Management Process 2008 ISO/IEC 12207 Software Lifecycle Processes 2009 ISO/IEC...1 Software And Systems Engineering Risk Management John Walz VP Technical and Conferences Activities, IEEE Computer Society Vice-Chair Planning...Software & Systems Engineering Standards Committee, IEEE Computer Society US TAG to ISO TMB Risk Management Working Group Systems and Software

  15. Multicore Considerations for Legacy Flight Software Migration

    Science.gov (United States)

    Vines, Kenneth; Day, Len

    2013-01-01

    In this paper we will discuss potential benefits and pitfalls when considering a migration from an existing single core code base to a multicore processor implementation. The results of this study present options that should be considered before migrating fault managers, device handlers and tasks with time-constrained requirements to a multicore flight software environment. Possible future multicore test bed demonstrations are also discussed.

  16. Software metrics: Software quality metrics for distributed systems. [reliability engineering

    Science.gov (United States)

    Post, J. V.

    1981-01-01

    Software quality metrics was extended to cover distributed computer systems. Emphasis is placed on studying embedded computer systems and on viewing them within a system life cycle. The hierarchy of quality factors, criteria, and metrics was maintained. New software quality factors were added, including survivability, expandability, and evolvability.

  17. A Layered Software Architecture for the Management of a Manufacturing Company

    Directory of Open Access Journals (Sweden)

    Domenico CONSOLI

    2011-01-01

    Full Text Available In this paper we describe a layered software architecture in the management of a manufactur-ing company that intensively uses computer technology. Application tools, new and legacy, after the updating, operate in a context of an open web oriented architecture. The software architecture enables the integration and interoperability among all tools that support business processes. Manufacturing Executive System and Text Mining tools are excellent interfaces, the former both for internal production and management processes and the latter for external processes coming from the market. In this way, it is possible to implement, a computer integrated factory, flexible and agile, that immediately responds to customer requirements.

  18. Osprey: Operating system for predictable clouds

    NARCIS (Netherlands)

    Sacha, Jan; Napper, Jeff; Mullender, Sape J.; McKie, Jim

    2012-01-01

    Cloud computing is currently based on hardware virtualization wherein a host operating system provides a virtual machine interface nearly identical to that of physical hardware to guest operating systems. Full transparency allows backward compatibility with legacy software but introduces

  19. Space Shuttle Program Primary Avionics Software System (PASS) Success Legacy -Major Accomplishments and Lessons Learned

    Science.gov (United States)

    Orr, James K.

    2010-01-01

    This presentation has shown the accomplishments of the PASS project over three decades and highlighted the lessons learned. Over the entire time, our goal has been to continuously improve our process, implement automation for both quality and increased productivity, and identify and remove all defects due to prior execution of a flawed process in addition to improving our processes following identification of significant process escapes. Morale and workforce instability have been issues, most significantly during 1993 to 1998 (period of consolidation in aerospace industry). The PASS project has also consulted with others, including the Software Engineering Institute, so as to be an early evaluator, adopter, and adapter of state-of-the-art software engineering innovations.

  20. Legacy to Industry 4.0: A Profibus Sniffer

    Science.gov (United States)

    Tsegaye Mamo, Fesseha; Sikora, Axel; Rathfelder, Christoph

    2017-07-01

    Legacy industrial communication protocols are proved robust and functional. During the last decades, the industry has invented completely new or advanced versions of the legacy communication solutions. However, even with the high adoption rate of these new solutions, still the majority industry applications run on legacy, mostly fieldbus related technologies. Profibus is one of those technologies that still keep on growing in the market, albeit a slow in market growth in recent years. A retrofit technology that would enable these technologies to connect to the Internet of Things, utilize the ever growing potential of data analysis, predictive maintenance or cloud-based application, while at the same time not changing a running system is fundamental.

  1. Software Quality Assurance for Nuclear Safety Systems

    International Nuclear Information System (INIS)

    Sparkman, D R; Lagdon, R

    2004-01-01

    The US Department of Energy has undertaken an initiative to improve the quality of software used to design and operate their nuclear facilities across the United States. One aspect of this initiative is to revise or create new directives and guides associated with quality practices for the safety software in its nuclear facilities. Safety software includes the safety structures, systems, and components software and firmware, support software and design and analysis software used to ensure the safety of the facility. DOE nuclear facilities are unique when compared to commercial nuclear or other industrial activities in terms of the types and quantities of hazards that must be controlled to protect workers, public and the environment. Because of these differences, DOE must develop an approach to software quality assurance that ensures appropriate risk mitigation by developing a framework of requirements that accomplishes the following goals: (sm b ullet) Ensures the software processes developed to address nuclear safety in design, operation, construction and maintenance of its facilities are safe (sm b ullet) Considers the larger system that uses the software and its impacts (sm b ullet) Ensures that the software failures do not create unsafe conditions Software designers for nuclear systems and processes must reduce risks in software applications by incorporating processes that recognize, detect, and mitigate software failure in safety related systems. It must also ensure that fail safe modes and component testing are incorporated into software design. For nuclear facilities, the consideration of risk is not necessarily sufficient to ensure safety. Systematic evaluation, independent verification and system safety analysis must be considered for software design, implementation, and operation. The software industry primarily uses risk analysis to determine the appropriate level of rigor applied to software practices. This risk-based approach distinguishes safety

  2. The economics of information systems and software

    CERN Document Server

    Veryard, Richard

    2014-01-01

    The Economics of Information Systems and Software focuses on the economic aspects of information systems and software, including advertising, evaluation of information systems, and software maintenance. The book first elaborates on value and values, software business, and scientific information as an economic category. Discussions focus on information products and information services, special economic properties of information, culture and convergence, hardware and software products, materiality and consumption, technological progress, and software flexibility. The text then takes a look at a

  3. Legacies in material flux: Structural changes before long-term studies

    Science.gov (United States)

    D.J. Bain; M.B. Green; J. Campbell; J. Chamblee; S. Chaoka; J. Fraterrigo; S. Kaushal; S. Martin; T. Jordan; T. Parolari; B. Sobczak; D. Weller; W. M. Wollheim; E. Boose; J. Duncan; G. Gettel; B. Hall; P. Kumar; J. Thompson; J. Vose; E. Elliott; D. Leigh

    2012-01-01

    Legacy effects of past land use and disturbance are increasingly recognized, yet consistent definitions of and criteria for defining them do not exist. To address this gap in biological- and ecosystem-assessment frameworks, we propose a general metric for evaluating potential legacy effects, which are computed by normalizing altered system function persistence with...

  4. Software for nuclear data acquisition systems

    International Nuclear Information System (INIS)

    Christensen, P.

    1983-01-01

    The situation for experimenters and system designers needing software for instrumentation is described. It is stated that software for a data acquisition system can be divided into programmes described as the foundation software, the applications programme, and the analysis programme. Special attention is given to CAMAC. Two examples from Risoe describing data transportation and archiving are given. Finally the supply of software and the problems of documentation are described. (author)

  5. Software Intensive Systems

    National Research Council Canada - National Science Library

    Horvitz, E; Katz, D. J; Rumpf, R. L; Shrobe, H; Smith, T. B; Webber, G. E; Williamson, W. E; Winston, P. H; Wolbarsht, James L

    2006-01-01

    .... Additionally, recommend that DoN invest in software engineering, particularly as it complements commercial industry developments and promotes the application of systems engineering methodology...

  6. Software Design Methods for Real-Time Systems

    Science.gov (United States)

    1989-12-01

    This module describes the concepts and methods used in the software design of real time systems . It outlines the characteristics of real time systems , describes...the role of software design in real time system development, surveys and compares some software design methods for real - time systems , and

  7. A Software Development Platform for Mechatronic Systems

    DEFF Research Database (Denmark)

    Guan, Wei

    Software has become increasingly determinative for development of mechatronic systems, which underscores the importance of demands for shortened time-to-market, increased productivity, higher quality, and improved dependability. As the complexity of systems is dramatically increasing, these demands...... present a challenge to the practitioners who adopt conventional software development approach. An effective approach towards industrial production of software for mechatronic systems is needed. This approach requires a disciplined engineering process that encompasses model-driven engineering and component......-based software engineering, whereby we enable incremental software development using component models to address the essential design issues of real-time embedded systems. To this end, this dissertation presents a software development platform that provides an incremental model-driven development process based...

  8. The Legacy of Nikola Tesla

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 12; Issue 4. The Legacy of Nikola Tesla - AC Power System and its Growth in India. D P Sen Gupta. General Article Volume 12 Issue 4 April 2007 pp 69-79. Fulltext. Click here to view fulltext PDF. Permanent link:

  9. Test software for BESIII MDC electronics system

    International Nuclear Information System (INIS)

    Zhang Hongyu; Sheng Huayi; Zhu Haitao; Ji Xiaolu; Zhao Dongxu

    2006-01-01

    This paper presents the design of Test System Software for BESIII MDC Electronics. Two kinds of test systems, SBS VP7 based and PowerPC based systems, and their corresponding test software are introduced. The software is developed in LabVIEW 7.1 and Microsoft Visual C++ 6.0, some test functions of the software, as well as their user interfaces, are described in detail. The software has been applied in hardware debugging, performance test and long term stability test. (authors)

  10. Statistical reliability assessment of software-based systems

    International Nuclear Information System (INIS)

    Korhonen, J.; Pulkkinen, U.; Haapanen, P.

    1997-01-01

    Plant vendors nowadays propose software-based systems even for the most critical safety functions. The reliability estimation of safety critical software-based systems is difficult since the conventional modeling techniques do not necessarily apply to the analysis of these systems, and the quantification seems to be impossible. Due to lack of operational experience and due to the nature of software faults, the conventional reliability estimation methods can not be applied. New methods are therefore needed for the safety assessment of software-based systems. In the research project Programmable automation systems in nuclear power plants (OHA), financed together by the Finnish Centre for Radiation and Nuclear Safety (STUK), the Ministry of Trade and Industry and the Technical Research Centre of Finland (VTT), various safety assessment methods and tools for software based systems are developed and evaluated. This volume in the OHA-report series deals with the statistical reliability assessment of software based systems on the basis of dynamic test results and qualitative evidence from the system design process. Other reports to be published later on in OHA-report series will handle the diversity requirements in safety critical software-based systems, generation of test data from operational profiles and handling of programmable automation in plant PSA-studies. (orig.) (25 refs.)

  11. The Legacy of Nikola Tesla

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 12; Issue 3. The Legacy of Nikola Tesla - The AC System that he Helped to Usher in. D P Sen Gupta. General Article Volume 12 Issue 3 March 2007 pp 54-69. Fulltext. Click here to view fulltext PDF. Permanent link:

  12. Analyzing Language in Suicide Notes and Legacy Tokens.

    Science.gov (United States)

    Egnoto, Michael J; Griffin, Darrin J

    2016-03-01

    Identifying precursors that will aid in the discovery of individuals who may harm themselves or others has long been a focus of scholarly research. This work set out to determine if it is possible to use the legacy tokens of active shooters and notes left from individuals who completed suicide to uncover signals that foreshadow their behavior. A total of 25 suicide notes and 21 legacy tokens were compared with a sample of over 20,000 student writings for a preliminary computer-assisted text analysis to determine what differences can be coded with existing computer software to better identify students who may commit self-harm or harm to others. The results support that text analysis techniques with the Linguistic Inquiry and Word Count (LIWC) tool are effective for identifying suicidal or homicidal writings as distinct from each other and from a variety of student writings in an automated fashion. Findings indicate support for automated identification of writings that were associated with harm to self, harm to others, and various other student writing products. This work begins to uncover the viability or larger scale, low cost methods of automatic detection for individuals suffering from harmful ideation.

  13. REDLetr: Workflow and tools to support the migration of legacy clinical data capture systems to REDCap.

    Science.gov (United States)

    Dunn, William D; Cobb, Jake; Levey, Allan I; Gutman, David A

    2016-09-01

    A memory clinic at an academic medical center has relied on several ad hoc data capture systems including Microsoft Access and Excel for cognitive assessments over the last several years. However these solutions are challenging to maintain and limit the potential of hypothesis-driven or longitudinal research. REDCap, a secure web application based on PHP and MySQL, is a practical solution for improving data capture and organization. Here, we present a workflow and toolset to facilitate legacy data migration and real-time clinical research data collection into REDCap as well as challenges encountered. Legacy data consisted of neuropsychological tests stored in over 4000 Excel workbooks. Functions for data extraction, norm scoring, converting to REDCap-compatible formats, accessing the REDCap API, and clinical report generation were developed and executed in Python. Over 400 unique data points for each workbook were migrated and integrated into our REDCap database. Moving forward, our REDCap-based system replaces the Excel-based data collection method as well as eases the integration into the standard clinical research workflow and Electronic Health Record. In the age of growing data, efficient organization and storage of clinical and research data is critical for advancing research and providing efficient patient care. We believe that the workflow and tools described in this work to promote legacy data integration as well as real time data collection into REDCap ultimately facilitate these goals. Published by Elsevier Ireland Ltd.

  14. Experimental research control software system

    International Nuclear Information System (INIS)

    Cohn, I A; Kovalenko, A G; Vystavkin, A N

    2014-01-01

    A software system, intended for automation of a small scale research, has been developed. The software allows one to control equipment, acquire and process data by means of simple scripts. The main purpose of that development is to increase experiment automation easiness, thus significantly reducing experimental setup automation efforts. In particular, minimal programming skills are required and supervisors have no reviewing troubles. Interactions between scripts and equipment are managed automatically, thus allowing to run multiple scripts simultaneously. Unlike well-known data acquisition commercial software systems, the control is performed by an imperative scripting language. This approach eases complex control and data acquisition algorithms implementation. A modular interface library performs interaction with external interfaces. While most widely used interfaces are already implemented, a simple framework is developed for fast implementations of new software and hardware interfaces. While the software is in continuous development with new features being implemented, it is already used in our laboratory for automation of a helium-3 cryostat control and data acquisition. The software is open source and distributed under Gnu Public License.

  15. Computer systems and software engineering

    Science.gov (United States)

    Mckay, Charles W.

    1988-01-01

    The High Technologies Laboratory (HTL) was established in the fall of 1982 at the University of Houston Clear Lake. Research conducted at the High Tech Lab is focused upon computer systems and software engineering. There is a strong emphasis on the interrelationship of these areas of technology and the United States' space program. In Jan. of 1987, NASA Headquarters announced the formation of its first research center dedicated to software engineering. Operated by the High Tech Lab, the Software Engineering Research Center (SERC) was formed at the University of Houston Clear Lake. The High Tech Lab/Software Engineering Research Center promotes cooperative research among government, industry, and academia to advance the edge-of-knowledge and the state-of-the-practice in key topics of computer systems and software engineering which are critical to NASA. The center also recommends appropriate actions, guidelines, standards, and policies to NASA in matters pertinent to the center's research. Results of the research conducted at the High Tech Lab/Software Engineering Research Center have given direction to many decisions made by NASA concerning the Space Station Program.

  16. Experimental research control software system

    Science.gov (United States)

    Cohn, I. A.; Kovalenko, A. G.; Vystavkin, A. N.

    2014-05-01

    A software system, intended for automation of a small scale research, has been developed. The software allows one to control equipment, acquire and process data by means of simple scripts. The main purpose of that development is to increase experiment automation easiness, thus significantly reducing experimental setup automation efforts. In particular, minimal programming skills are required and supervisors have no reviewing troubles. Interactions between scripts and equipment are managed automatically, thus allowing to run multiple scripts simultaneously. Unlike well-known data acquisition commercial software systems, the control is performed by an imperative scripting language. This approach eases complex control and data acquisition algorithms implementation. A modular interface library performs interaction with external interfaces. While most widely used interfaces are already implemented, a simple framework is developed for fast implementations of new software and hardware interfaces. While the software is in continuous development with new features being implemented, it is already used in our laboratory for automation of a helium-3 cryostat control and data acquisition. The software is open source and distributed under Gnu Public License.

  17. Automating software design system DESTA

    Science.gov (United States)

    Lovitsky, Vladimir A.; Pearce, Patricia D.

    1992-01-01

    'DESTA' is the acronym for the Dialogue Evolutionary Synthesizer of Turnkey Algorithms by means of a natural language (Russian or English) functional specification of algorithms or software being developed. DESTA represents the computer-aided and/or automatic artificial intelligence 'forgiving' system which provides users with software tools support for algorithm and/or structured program development. The DESTA system is intended to provide support for the higher levels and earlier stages of engineering design of software in contrast to conventional Computer Aided Design (CAD) systems which provide low level tools for use at a stage when the major planning and structuring decisions have already been taken. DESTA is a knowledge-intensive system. The main features of the knowledge are procedures, functions, modules, operating system commands, batch files, their natural language specifications, and their interlinks. The specific domain for the DESTA system is a high level programming language like Turbo Pascal 6.0. The DESTA system is operational and runs on an IBM PC computer.

  18. Requirements engineering for software and systems

    CERN Document Server

    Laplante, Phillip A

    2014-01-01

    Solid requirements engineering has increasingly been recognized as the key to improved, on-time and on-budget delivery of software and systems projects. This book provides practical teaching for graduate and professional systems and software engineers. It uses extensive case studies and exercises to help students grasp concepts and techniques. With a focus on software-intensive systems, this text provides a probing and comprehensive review of recent developments in intelligent systems, soft computing techniques, and their diverse applications in manufacturing. The second edition contains 100% revised content and approximately 30% new material

  19. Virtual Exercise Training Software System

    Science.gov (United States)

    Vu, L.; Kim, H.; Benson, E.; Amonette, W. E.; Barrera, J.; Perera, J.; Rajulu, S.; Hanson, A.

    2018-01-01

    The purpose of this study was to develop and evaluate a virtual exercise training software system (VETSS) capable of providing real-time instruction and exercise feedback during exploration missions. A resistive exercise instructional system was developed using a Microsoft Kinect depth-camera device, which provides markerless 3-D whole-body motion capture at a small form factor and minimal setup effort. It was hypothesized that subjects using the newly developed instructional software tool would perform the deadlift exercise with more optimal kinematics and consistent technique than those without the instructional software. Following a comprehensive evaluation in the laboratory, the system was deployed for testing and refinement in the NASA Extreme Environment Mission Operations (NEEMO) analog.

  20. REENGINEERING OF THE AIR SIMULATORS LEGACY SOFTWARE

    Directory of Open Access Journals (Sweden)

    Nikolay O. Sidorov

    2008-02-01

    Full Text Available  There are the technical complexes consisting of components, parts of which are actively used, but the rest has lost working capacity owing to moral and physical deterioration. An example of such a complex is the aviation-flight complex "plane-simulator". High cost of components which continue to be used (plane do the actual task of restoring and supporting the out-of-order components (simulator. The considerable part of such complexes is the software, which owing to replacement of the obsolete and physically worn out hardware requires the rework. The rework method is reengineering.

  1. MAST data acquisition system

    International Nuclear Information System (INIS)

    Shibaev, S.; Counsell, G.; Cunningham, G.; Manhood, S.J.; Thomas-Davies, N.; Waterhouse, J.

    2006-01-01

    The data acquisition system of the Mega-Amp Spherical Tokamak (MAST) presently collects up to 400 MB of data in about 3000 data items per shot, and subsequent fast growth is expected. Since the start of MAST operations (in 1999) the system has changed dramatically. Though we continue to use legacy CAMAC hardware, newer VME, PCI, and PXI based sub-systems collect most of the data now. All legacy software has been redesigned and new software has been developed. Last year a major system improvement was made-replacement of the message distribution system. The new message system provides easy connection of any sub-system independently of its platform and serves as a framework for many new applications. A new data acquisition controller provides full control of common sub-systems, central error logging, and data acquisition alarms for the MAST plant. A number of new sub-systems using Linux and Windows OSs on VME, PCI, and PXI platforms have been developed. A new PXI unit has been designed as a base sub-system accommodating any type of data acquisition and control devices. Several web applications for the real-time MAST monitoring and data presentation have been developed

  2. The legacy of pesticide pollution: An overlooked factor in current risk assessments of freshwater systems

    DEFF Research Database (Denmark)

    Rasmussen, Jes Jessen; Wiberg-Larsen, Peter; Baattrup-Pedersen, Annette

    2015-01-01

    We revealed a history of legacy pesticides in water and sediment samples from 19 small streams across an agricultural landscape. Dominant legacy compounds included organochlorine pesticides, such as DDT and lindane, the organophosphate chlorpyrifos and triazine herbicides such as terbutylazine...

  3. Certification of digital system software

    International Nuclear Information System (INIS)

    Waclo, J.; Cook, B.; Adomaitis, D.

    1991-01-01

    The activities involved in the successful application of digital systems to Nuclear Protection functions is not achieved through happenstance. At Westinghouse there has been a longstanding program to utilize digital state of the art technology for protection system advancement. Thereby gaining the advantages of increased system reliability, performance, ease of operation and reduced maintenance costs. This paper describes the Westinghouse background and experience in the safety system software development process, including Verification and Validation, and its application to protection system qualification and the successful use for licensing the Eagle 21 Digital Process Protection System Upgrade. In addition, the lessons learned from this experience are discussed from the perspective of improving the development process through applying feedback of the measurements made on the process and the software product quality. The goal of this process optimization is to produce the highest possible software quality while recognizing the real world constraints of available resources, project schedule and the regulatory policies that are customary in the nuclear industry

  4. A study of software safety analysis system for safety-critical software

    International Nuclear Information System (INIS)

    Chang, H. S.; Shin, H. K.; Chang, Y. W.; Jung, J. C.; Kim, J. H.; Han, H. H.; Son, H. S.

    2004-01-01

    The core factors and requirements for the safety-critical software traced and the methodology adopted in each stage of software life cycle are presented. In concept phase, Failure Modes and Effects Analysis (FMEA) for the system has been performed. The feasibility evaluation of selected safety parameter was performed and Preliminary Hazards Analysis list was prepared using HAZOP(Hazard and Operability) technique. And the check list for management control has been produced via walk-through technique. Based on the evaluation of the check list, activities to be performed in requirement phase have been determined. In the design phase, hazard analysis has been performed to check the safety capability of the system with regard to safety software algorithm using Fault Tree Analysis (FTA). In the test phase, the test items based on FMEA have been checked for fitness guided by an accident scenario. The pressurizer low pressure trip algorithm has been selected to apply FTA method to software safety analysis as a sample. By applying CASE tool, the requirements traceability of safety critical system has been enhanced during all of software life cycle phases

  5. Repurposing legacy data innovative case studies

    CERN Document Server

    Berman, Jules J

    2015-01-01

    Repurposing Legacy Data: Innovative Case Studies takes a look at how data scientists have re-purposed legacy data, whether their own, or legacy data that has been donated to the public domain. Most of the data stored worldwide is legacy data-data created some time in the past, for a particular purpose, and left in obsolete formats. As with keepsakes in an attic, we retain this information thinking it may have value in the future, though we have no current use for it. The case studies in this book, from such diverse fields as cosmology, quantum physics, high-energy physics, microbiology,

  6. An Architecture, System Engineering, and Acquisition Approach for Space System Software Resiliency

    Science.gov (United States)

    Phillips, Dewanne Marie

    Software intensive space systems can harbor defects and vulnerabilities that may enable external adversaries or malicious insiders to disrupt or disable system functions, risking mission compromise or loss. Mitigating this risk demands a sustained focus on the security and resiliency of the system architecture including software, hardware, and other components. Robust software engineering practices contribute to the foundation of a resilient system so that the system "can take a hit to a critical component and recover in a known, bounded, and generally acceptable period of time". Software resiliency must be a priority and addressed early in the life cycle development to contribute a secure and dependable space system. Those who develop, implement, and operate software intensive space systems must determine the factors and systems engineering practices to address when investing in software resiliency. This dissertation offers methodical approaches for improving space system resiliency through software architecture design, system engineering, increased software security, thereby reducing the risk of latent software defects and vulnerabilities. By providing greater attention to the early life cycle phases of development, we can alter the engineering process to help detect, eliminate, and avoid vulnerabilities before space systems are delivered. To achieve this objective, this dissertation will identify knowledge, techniques, and tools that engineers and managers can utilize to help them recognize how vulnerabilities are produced and discovered so that they can learn to circumvent them in future efforts. We conducted a systematic review of existing architectural practices, standards, security and coding practices, various threats, defects, and vulnerabilities that impact space systems from hundreds of relevant publications and interviews of subject matter experts. We expanded on the system-level body of knowledge for resiliency and identified a new software

  7. Altmetrics, Legacy Scholarship, and Scholarly Legacy

    Directory of Open Access Journals (Sweden)

    Lauren B. Collister

    2017-10-01

    Full Text Available When using alternative metrics (altmetrics to investigate the impact of a scholar’s work, researchers and librarians are typically cautioned that altmetrics will be less useful for older works of scholarship. This is because it is difficult to collect social media and other attention retroactively, and the numbers will be lower if the work was published before social media marketing and promotion were widely accepted in a field. In this article, we argue that altmetrics can provide useful information about older works in the form of documenting renewed attention to past scholarship as part of a scholar’s legacy. Using the altmetrics profile of the late Dr. Thomas E. Starzl, often referred to as “the father of modern transplantation”, we describe two cases where altmetrics provided information about renewed interest in his works: a controversy about race and genetics that shows the ongoing impact of a particular work, and posthumous remembrances by colleagues which reveal his scholarly legacy.

  8. Legacy material work-off project

    International Nuclear Information System (INIS)

    Sloan, T.J.; Baker, D.H. IV

    1999-01-01

    Los Alamos National Laboratory (LANL) and its subcontractors recently completed a nine-month legacy material clean-up effort. Legacy materials were defined as chemicals, hazardous, non-hazardous, and both hazardous and radioactive (mixed), that no longer served a programmatic use and had no identified individual owner within the Laboratory. Once personnel identified the legacy materials, the items were transferred to Solid Waste Operation's (EM-SWO) control. Upon completing this process, the responsible division-level manager was required to certify that all non-radioactive hazardous and non-hazardous materials and acceptable mixed legacy materials had been identified and transferred to EM-SWO for proper handling or disposal. The major expense in this project was the cost of actual chemical and radiological analysis. This expense was the result of items not having an identified individual owner. The major benefit of this project is that LANL is now in an excellent position to implement its Integrated Safety Management (ISM) Plan, which requires the implementation of safe work practices, including requirements for removing unused items when vacating workspaces. Effective implementation of ISM will go a long way toward ensuring that legacy materials are no longer an issue at the Laboratory

  9. Flight test of a resident backup software system

    Science.gov (United States)

    Deets, Dwain A.; Lock, Wilton P.; Megna, Vincent A.

    1987-01-01

    A new fault-tolerant system software concept employing the primary digital computers as host for the backup software portion has been implemented and flight tested in the F-8 digital fly-by-wire airplane. The system was implemented in such a way that essentially no transients occurred in transferring from primary to backup software. This was accomplished without a significant increase in the complexity of the backup software. The primary digital system was frame synchronized, which provided several advantages in implementing the resident backup software system. Since the time of the flight tests, two other flight vehicle programs have made a commitment to incorporate resident backup software similar in nature to the system described here.

  10. Tools for Embedded Computing Systems Software

    Science.gov (United States)

    1978-01-01

    A workshop was held to assess the state of tools for embedded systems software and to determine directions for tool development. A synopsis of the talk and the key figures of each workshop presentation, together with chairmen summaries, are presented. The presentations covered four major areas: (1) tools and the software environment (development and testing); (2) tools and software requirements, design, and specification; (3) tools and language processors; and (4) tools and verification and validation (analysis and testing). The utility and contribution of existing tools and research results for the development and testing of embedded computing systems software are described and assessed.

  11. Software design for resilient computer systems

    CERN Document Server

    Schagaev, Igor

    2016-01-01

    This book addresses the question of how system software should be designed to account for faults, and which fault tolerance features it should provide for highest reliability. The authors first show how the system software interacts with the hardware to tolerate faults. They analyze and further develop the theory of fault tolerance to understand the different ways to increase the reliability of a system, with special attention on the role of system software in this process. They further develop the general algorithm of fault tolerance (GAFT) with its three main processes: hardware checking, preparation for recovery, and the recovery procedure. For each of the three processes, they analyze the requirements and properties theoretically and give possible implementation scenarios and system software support required. Based on the theoretical results, the authors derive an Oberon-based programming language with direct support of the three processes of GAFT. In the last part of this book, they introduce a simulator...

  12. Software engineering practices for control system reliability

    International Nuclear Information System (INIS)

    S. K. Schaffner; K. S White

    1999-01-01

    This paper will discuss software engineering practices used to improve Control System reliability. The authors begin with a brief discussion of the Software Engineering Institute's Capability Maturity Model (CMM) which is a framework for evaluating and improving key practices used to enhance software development and maintenance capabilities. The software engineering processes developed and used by the Controls Group at the Thomas Jefferson National Accelerator Facility (Jefferson Lab), using the Experimental Physics and Industrial Control System (EPICS) for accelerator control, are described. Examples are given of how their procedures have been used to minimized control system downtime and improve reliability. While their examples are primarily drawn from their experience with EPICS, these practices are equally applicable to any control system. Specific issues addressed include resource allocation, developing reliable software lifecycle processes and risk management

  13. Software dependability in the Tandem GUARDIAN system

    Science.gov (United States)

    Lee, Inhwan; Iyer, Ravishankar K.

    1995-01-01

    Based on extensive field failure data for Tandem's GUARDIAN operating system this paper discusses evaluation of the dependability of operational software. Software faults considered are major defects that result in processor failures and invoke backup processes to take over. The paper categorizes the underlying causes of software failures and evaluates the effectiveness of the process pair technique in tolerating software faults. A model to describe the impact of software faults on the reliability of an overall system is proposed. The model is used to evaluate the significance of key factors that determine software dependability and to identify areas for improvement. An analysis of the data shows that about 77% of processor failures that are initially considered due to software are confirmed as software problems. The analysis shows that the use of process pairs to provide checkpointing and restart (originally intended for tolerating hardware faults) allows the system to tolerate about 75% of reported software faults that result in processor failures. The loose coupling between processors, which results in the backup execution (the processor state and the sequence of events) being different from the original execution, is a major reason for the measured software fault tolerance. Over two-thirds (72%) of measured software failures are recurrences of previously reported faults. Modeling, based on the data, shows that, in addition to reducing the number of software faults, software dependability can be enhanced by reducing the recurrence rate.

  14. Honeywell Modular Automation System Computer Software Documentation

    International Nuclear Information System (INIS)

    STUBBS, A.M.

    2000-01-01

    The purpose of this Computer Software Document (CSWD) is to provide configuration control of the Honeywell Modular Automation System (MAS) in use at the Plutonium Finishing Plant (PFP). This CSWD describes hardware and PFP developed software for control of stabilization furnaces. The Honeywell software can generate configuration reports for the developed control software. These reports are described in the following section and are attached as addendum's. This plan applies to PFP Engineering Manager, Thermal Stabilization Cognizant Engineers, and the Shift Technical Advisors responsible for the Honeywell MAS software/hardware and administration of the Honeywell System

  15. Simulation software support (S3) system a software testing and debugging tool

    International Nuclear Information System (INIS)

    Burgess, D.C.; Mahjouri, F.S.

    1990-01-01

    The largest percentage of technical effort in the software development process is accounted for debugging and testing. It is not unusual for a software development organization to spend over 50% of the total project effort on testing. In the extreme, testing of human-rated software (e.g., nuclear reactor monitoring, training simulator) can cost three to five times as much as all other software engineering steps combined. The Simulation Software Support (S 3 ) System, developed by the Link-Miles Simulation Corporation is ideally suited for real-time simulation applications which involve a large database with models programmed in FORTRAN. This paper will focus on testing elements of the S 3 system. In this paper system support software utilities are provided which enable the loading and execution of modules in the development environment. These elements include the Linking/Loader (LLD) for dynamically linking program modules and loading them into memory and the interactive executive (IEXEC) for controlling the execution of the modules. Features of the Interactive Symbolic Debugger (SD) and the Real Time Executive (RTEXEC) to support the unit and integrated testing will be explored

  16. SWATCH Common software for controlling and monitoring the upgraded CMS Level-1 trigger

    CERN Document Server

    Lazaridis, Christos; Bunkowski, Karol; Codispoti, Giuseppe; Dirkx, Glenn; Ghabrous Larrea, Carlos; Lingemann, Joschka; Kreczko, Lukasz; Thea, Alessandro; Williams, Tom

    2017-01-01

    The Large Hadron Collider at CERN restarted in 2015 with a higher centre-of-mass energy of 13 TeV. The instantaneous luminosity is expected to increase significantly in the coming years. An upgraded Level-1 trigger system is being deployed in the CMS experiment in order to maintain the same efficiencies for searches and precision measurements as those achieved in the previous run. This system must be controlled and monitored coherently through software, with high operational efficiency.The legacy system is composed of approximately 4000 data processor boards, of several custom application-specific designs. These boards are organised into several subsystems; each subsystem receives data from different detector systems (calorimeters, barrel/endcap muon detectors), or with differing granularity. These boards have been controlled and monitored by a medium-sized distributed system of over 40 computers and 200 processes. Only a small fraction of the control and monitoring software was common between the different s...

  17. Observation-Driven Configuration of Complex Software Systems

    Science.gov (United States)

    Sage, Aled

    2010-06-01

    The ever-increasing complexity of software systems makes them hard to comprehend, predict and tune due to emergent properties and non-deterministic behaviour. Complexity arises from the size of software systems and the wide variety of possible operating environments: the increasing choice of platforms and communication policies leads to ever more complex performance characteristics. In addition, software systems exhibit different behaviour under different workloads. Many software systems are designed to be configurable so that policies can be chosen to meet the needs of various stakeholders. For complex software systems it can be difficult to accurately predict the effects of a change and to know which configuration is most appropriate. This thesis demonstrates that it is useful to run automated experiments that measure a selection of system configurations. Experiments can find configurations that meet the stakeholders' needs, find interesting behavioural characteristics, and help produce predictive models of the system's behaviour. The design and use of ACT (Automated Configuration Tool) for running such experiments is described, in combination a number of search strategies for deciding on the configurations to measure. Design Of Experiments (DOE) is discussed, with emphasis on Taguchi Methods. These statistical methods have been used extensively in manufacturing, but have not previously been used for configuring software systems. The novel contribution here is an industrial case study, applying the combination of ACT and Taguchi Methods to DC-Directory, a product from Data Connection Ltd (DCL). The case study investigated the applicability of Taguchi Methods for configuring complex software systems. Taguchi Methods were found to be useful for modelling and configuring DC- Directory, making them a valuable addition to the techniques available to system administrators and developers.

  18. 14 CFR 417.123 - Computing systems and software.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 4 2010-01-01 2010-01-01 false Computing systems and software. 417.123... systems and software. (a) A launch operator must document a system safety process that identifies the... systems and software. (b) A launch operator must identify all safety-critical functions associated with...

  19. Legacy effects in material flux: structural catchment changes predate long-term studies

    Science.gov (United States)

    Daniel Bain; Mark B. Green; John L. Campbell; John F. Chamblee; Sayo Chaoka; Jennifer M. Fraterrigo; Sujay S. Kaushal; Sujay S. Kaushal; Sherry L. Martin; Thomas E. Jordan; Anthony J. Parolari; William V. Sobczak; Donald E. Weller; Wilfred M. Wolheim; Emery R. Boose; Jonathan M. Duncan; Gretchen M. Gettel; Brian R. Hall; Praveen Kumar; Jonathan R. Thompson; James M. Vose; Emily M. Elliott; David S. Leigh

    2012-01-01

    Legacy effects of past land use and disturbance are increasingly recognized, yet consistent definitions of and criteria for defining them do not exist. To address this gap in biological- and ecosystem-assessment frameworks, we propose a general metric for evaluating potential legacy effects, which are computed by normalizing altered system function persistence with...

  20. Resilience Engineering in Critical Long Term Aerospace Software Systems: A New Approach to Spacecraft Software Safety

    Science.gov (United States)

    Dulo, D. A.

    Safety critical software systems permeate spacecraft, and in a long term venture like a starship would be pervasive in every system of the spacecraft. Yet software failure today continues to plague both the systems and the organizations that develop them resulting in the loss of life, time, money, and valuable system platforms. A starship cannot afford this type of software failure in long journeys away from home. A single software failure could have catastrophic results for the spaceship and the crew onboard. This paper will offer a new approach to developing safe reliable software systems through focusing not on the traditional safety/reliability engineering paradigms but rather by focusing on a new paradigm: Resilience and Failure Obviation Engineering. The foremost objective of this approach is the obviation of failure, coupled with the ability of a software system to prevent or adapt to complex changing conditions in real time as a safety valve should failure occur to ensure safe system continuity. Through this approach, safety is ensured through foresight to anticipate failure and to adapt to risk in real time before failure occurs. In a starship, this type of software engineering is vital. Through software developed in a resilient manner, a starship would have reduced or eliminated software failure, and would have the ability to rapidly adapt should a software system become unstable or unsafe. As a result, long term software safety, reliability, and resilience would be present for a successful long term starship mission.

  1. SCaN Testbed Software Development and Lessons Learned

    Science.gov (United States)

    Kacpura, Thomas J.; Varga, Denise M.

    2012-01-01

    National Aeronautics and Space Administration (NASA) has developed an on-orbit, adaptable, Software Defined Radio (SDR)Space Telecommunications Radio System (STRS)-based testbed facility to conduct a suite of experiments to advance technologies, reduce risk, and enable future mission capabilities on the International Space Station (ISS). The SCAN Testbed Project will provide NASA, industry, other Government agencies, and academic partners the opportunity to develop and field communications, navigation, and networking technologies in the laboratory and space environment based on reconfigurable, SDR platforms and the STRS Architecture.The SDRs are a new technology for NASA, and the support infrastructure they require is different from legacy, fixed function radios. SDRs offer the ability to reconfigure on-orbit communications by changing software for new waveforms and operating systems to enable new capabilities or fix any anomalies, which was not a previous option. They are not stand alone devices, but required a new approach to effectively control them and flow data. This requires extensive software to be developed to utilize the full potential of these reconfigurable platforms. The paper focuses on development, integration and testing as related to the avionics processor system, and the software required to command, control, monitor, and interact with the SDRs, as well as the other communication payload elements. An extensive effort was required to develop the flight software and meet the NASA requirements for software quality and safety. The flight avionics must be radiation tolerant, and these processors have limited capability in comparison to terrestrial counterparts. A big challenge was that there are three SDRs onboard, and interfacing with multiple SDRs simultaneously complicatesd the effort. The effort also includes ground software, which is a key element for both the command of the payload, and displaying data created by the payload. The verification of

  2. Space Telescope Pointing Control System software

    Science.gov (United States)

    Dougherty, H.; Rodoni, C.; Rossini, R.; Tompetrini, K.; Nakashima, A.; Bradley, A.

    1982-01-01

    The Space Telescope Pointing Control System software is in the advanced development stage, having been tested on both the airbearing and the static simulator. The overall structure of the software is discussed, along with timing and sizing evaluations. The interaction between the controls analysts and software designer is described.

  3. A measurement system for large, complex software programs

    Science.gov (United States)

    Rone, Kyle Y.; Olson, Kitty M.; Davis, Nathan E.

    1994-01-01

    This paper describes measurement systems required to forecast, measure, and control activities for large, complex software development and support programs. Initial software cost and quality analysis provides the foundation for meaningful management decisions as a project evolves. In modeling the cost and quality of software systems, the relationship between the functionality, quality, cost, and schedule of the product must be considered. This explicit relationship is dictated by the criticality of the software being developed. This balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and developers with respect to the processes being employed.

  4. U.S. Department of Energy Office of Legacy Management Legacy Uranium Mine Site Reclamation - Lessons Learned - 12384

    Energy Technology Data Exchange (ETDEWEB)

    Kilpatrick, Laura E. [U.S. Department of Energy Office of Legacy Management, Westminster, Colorado 80021 (United States); Cotter, Ed [S.M. Stoller Corporation, Grand Junction, Colorado 81503 (United States)

    2012-07-01

    The U.S. Department of Energy (DOE) Office of Legacy Management is responsible for administering the DOE Uranium Leasing Program (ULP) and its 31 uranium lease tracts located in the Uravan Mineral Belt of southwestern Colorado (see Figure 1). In addition to administering the ULP for the last six decades, DOE has also undertaken the significant task of reclaiming a large number of abandoned uranium (legacy) mine sites and associated features located throughout the Uravan Mineral Belt. In 1995, DOE initiated a 3-year reconnaissance program to locate and delineate (through extensive on-the-ground mapping) the legacy mine sites and associated features contained within the historically defined boundaries of its uranium lease tracts. During that same time frame, DOE recognized the lack of regulations pertaining to the reclamation of legacy mine sites and contacted the U.S. Bureau of Land Management (BLM) concerning the reclamation of legacy mine sites. In November 1995, The BLM Colorado State Office formally issued the United States Department of the Interior, Colorado Bureau of Land Management, Closure/Reclamation Guidelines, Abandoned Uranium Mine Sites as a supplement to its Solid Minerals Reclamation Handbook (H-3042-1). Over the next five-and-one-half years, DOE reclaimed the 161 legacy mine sites that had been identified on DOE withdrawn lands. By the late 1990's, the various BLM field offices in southwestern Colorado began to recognize DOE's experience and expertise in reclaiming legacy mine sites. During the ensuing 8 years, BLM funded DOE (through a series of task orders) to perform reclamation activities at 182 BLM mine sites. To date, DOE has reclaimed 372 separate and distinct legacy mine sites. During this process, DOE has learned many lessons and is willing to share those lessons with others in the reclamation industry because there are still many legacy mine sites not yet reclaimed. DOE currently administers 31 lease tracts (11,017 ha) that

  5. Author's Response to Commentaries on: "An Interpretation of Part of Gilbert Gottlieb's Legacy: Developmental Systems Theory Contra Developmental Behavior Genetics"

    Science.gov (United States)

    Molenaar, Peter C. M.

    2015-01-01

    In this article, Peter Molenaar responds to three commentaries (this issue) on his article, "An Interpretation of Part of Gilbert Gottlieb's Legacy: Developmental Systems Theory Contra Developmental Behavior Genetics." He addresses aspects of relational developmental systems (RDS) mentioned and questions raised in each of the…

  6. Software Engineering Issues for Cyber-Physical Systems

    DEFF Research Database (Denmark)

    Al-Jaroodi, Jameela; Mohamed, Nader; Jawhar, Imad

    2016-01-01

    step; however, designing and implementing the right software to integrate and use them effectively is essential. The software facilitates better interfaces, more control and adds smart services, high flexibility and many other added values and features to the CPS. However, software development for CPS......Cyber-Physical Systems (CPS) provide many smart features for enhancing physical processes. These systems are designed with a set of distributed hardware, software, and network components that are embedded in physical systems and environments or attached to humans. Together they function seamlessly...... to offer specific functionalities or features that help enhance human lives, operations or environments. While different CPS components play important roles in a successful CPS development, the software plays the most important role among them. Acquiring and using high quality CPS components is the first...

  7. Distributed Storage Inverter and Legacy Generator Integration Plus Renewables Solution for Microgrids

    Science.gov (United States)

    2015-07-01

    inverter (advantage of short term storage) and enables microgrid upgrade of legacy generator assets (integration of inverter and generator controllers ...today. The storage inverter controls and hardware are also leveraged to provide an islanding inverter ( microgrid compatibility) for renewable...transiently rated inverters, integration with legacy generator controls , and microgrid compatible inverters for PV. Key system design drivers are lowering

  8. Interaction between systems and software engineering in safety-critical systems

    International Nuclear Information System (INIS)

    Knight, J.

    1994-01-01

    There are three areas of concern: when is software to be considered safe; what, exactly, is the role of the software engineer; and how do systems, or sometimes applications, engineers and software engineers interact with each other. The author presents his perspective on these questions which he feels differ from those of many in the field. He argues for a clear definition of safety in the software arena, so the engineer knows what he is engineering toward. Software must be viewed as part of the entire system, since it does not function on its own, or isolation. He argues for the establishment of clear specifications in this area

  9. 14 CFR 415.123 - Computing systems and software.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 4 2010-01-01 2010-01-01 false Computing systems and software. 415.123... Launch Vehicle From a Non-Federal Launch Site § 415.123 Computing systems and software. (a) An applicant's safety review document must describe all computing systems and software that perform a safety...

  10. The software product assurance metrics study: JPL's software systems quality and productivity

    Science.gov (United States)

    Bush, Marilyn W.

    1989-01-01

    The findings are reported of the Jet Propulsion Laboratory (JPL)/Software Product Assurance (SPA) Metrics Study, conducted as part of a larger JPL effort to improve software quality and productivity. Until recently, no comprehensive data had been assembled on how JPL manages and develops software-intensive systems. The first objective was to collect data on software development from as many projects and for as many years as possible. Results from five projects are discussed. These results reflect 15 years of JPL software development, representing over 100 data points (systems and subsystems), over a third of a billion dollars, over four million lines of code and 28,000 person months. Analysis of this data provides a benchmark for gauging the effectiveness of past, present and future software development work. In addition, the study is meant to encourage projects to record existing metrics data and to gather future data. The SPA long term goal is to integrate the collection of historical data and ongoing project data with future project estimations.

  11. Data systems and computer science: Software Engineering Program

    Science.gov (United States)

    Zygielbaum, Arthur I.

    1991-01-01

    An external review of the Integrated Technology Plan for the Civil Space Program is presented. This review is specifically concerned with the Software Engineering Program. The goals of the Software Engineering Program are as follows: (1) improve NASA's ability to manage development, operation, and maintenance of complex software systems; (2) decrease NASA's cost and risk in engineering complex software systems; and (3) provide technology to assure safety and reliability of software in mission critical applications.

  12. A NEW EXHAUST VENTILATION SYSTEM DESIGN SOFTWARE

    Directory of Open Access Journals (Sweden)

    H. Asilian Mahabady

    2007-09-01

    Full Text Available A Microsoft Windows based ventilation software package is developed to reduce time-consuming and boring procedure of exhaust ventilation system design. This program Assure accurate and reliable air pollution control related calculations. Herein, package is tentatively named Exhaust Ventilation Design Software which is developed in VB6 programming environment. Most important features of Exhaust Ventilation Design Software that are ignored in formerly developed packages are Collector design and fan dimension data calculations. Automatic system balance is another feature of this package. Exhaust Ventilation Design Software algorithm for design is based on two methods: Balance by design (Static pressure balance and design by Blast gate. The most important section of software is a spreadsheet that is designed based on American Conference of Governmental Industrial Hygienists calculation sheets. Exhaust Ventilation Design Software is developed so that engineers familiar with American Conference of Governmental Industrial Hygienists datasheet can easily employ it for ventilation systems design. Other sections include Collector design section (settling chamber, cyclone, and packed tower, fan geometry and dimension data section, a unit converter section (that helps engineers to deal with units, a hood design section and a Persian HTML help. Psychometric correction is also considered in Exhaust Ventilation Design Software. In Exhaust Ventilation Design Software design process, efforts are focused on improving GUI (graphical user interface and use of programming standards in software design. Reliability of software has been evaluated and results show acceptable accuracy.

  13. Software qualification for digital safety system in KNICS project

    International Nuclear Information System (INIS)

    Kwon, Kee-Choon; Lee, Dong-Young; Choi, Jong-Gyun

    2012-01-01

    In order to achieve technical self-reliance in the area of nuclear instrumentation and control, the Korea Nuclear Instrumentation and Control System (KNICS) project had been running for seven years from 2001. The safety-grade Programmable Logic Controller (PLC) and the digital safety system were developed by KNICS project. All the software of the PLC and digital safety system were developed and verified following the software development life cycle Verification and Validation (V and V) procedure. The main activities of the V and V process are preparation of software planning documentations, verification of the Software Requirement Specification (SRS), Software Design Specification (SDS) and codes, and a testing of the software components, the integrated software, and the integrated system. In addition, a software safety analysis and a software configuration management are included in the activities. For the software safety analysis at the SRS and SDS phases, the software Hazard Operability (HAZOP) was performed and then the software fault tree analysis was applied. The software fault tree analysis was applied to a part of software module with some critical defects identified by the software HAZOP in SDS phase. The software configuration management was performed using the in-house tool developed in the KNICS project. (author)

  14. A fault-tolerant software strategy for digital systems

    Science.gov (United States)

    Hitt, E. F.; Webb, J. J.

    1984-01-01

    Techniques developed for producing fault-tolerant software are described. Tolerance is required because of the impossibility of defining fault-free software. Faults are caused by humans and can appear anywhere in the software life cycle. Tolerance is effected through error detection, damage assessment, recovery, and fault treatment, followed by return of the system to service. Multiversion software comprises two or more versions of the software yielding solutions which are examined by a decision algorithm. Errors can also be detected by extrapolation from previous results or by the acceptability of results. Violations of timing specifications can reveal errors, or the system can roll back to an error-free state when a defect is detected. The software, when used in flight control systems, must not impinge on time-critical responses. Efforts are still needed to reduce the costs of developing the fault-tolerant systems.

  15. Expert System Software Assistant for Payload Operations

    Science.gov (United States)

    Rogers, Mark N.

    1997-01-01

    The broad objective of this expert system software based application was to demonstrate the enhancements and cost savings that can be achieved through expert system software utilization in a spacecraft ground control center. Spacelab provided a valuable proving ground for this advanced software technology; a technology that will be exploited and expanded for future ISS operations. Our specific focus was on demonstrating payload cadre command and control efficiency improvements through the use of "smart" software which monitors flight telemetry, provides enhanced schematic-based data visualization, and performs advanced engineering data analysis.

  16. Trend Monitoring System (TMS) graphics software

    Science.gov (United States)

    Brown, J. S.

    1979-01-01

    A prototype bus communications systems, which is being used to support the Trend Monitoring System (TMS) and to evaluate the bus concept is considered. A set of FORTRAN-callable graphics subroutines for the host MODCOMP comuter, and an approach to splitting graphics work between the host and the system's intelligent graphics terminals are described. The graphics software in the MODCOMP and the operating software package written for the graphics terminals are included.

  17. RT-Syn: A real-time software system generator

    Science.gov (United States)

    Setliff, Dorothy E.

    1992-01-01

    This paper presents research into providing highly reusable and maintainable components by using automatic software synthesis techniques. This proposal uses domain knowledge combined with automatic software synthesis techniques to engineer large-scale mission-critical real-time software. The hypothesis centers on a software synthesis architecture that specifically incorporates application-specific (in this case real-time) knowledge. This architecture synthesizes complex system software to meet a behavioral specification and external interaction design constraints. Some examples of these external constraints are communication protocols, precisions, timing, and space limitations. The incorporation of application-specific knowledge facilitates the generation of mathematical software metrics which are used to narrow the design space, thereby making software synthesis tractable. Success has the potential to dramatically reduce mission-critical system life-cycle costs not only by reducing development time, but more importantly facilitating maintenance, modifications, and extensions of complex mission-critical software systems, which are currently dominating life cycle costs.

  18. Analyzing Software Requirements Errors in Safety-Critical, Embedded Systems

    Science.gov (United States)

    Lutz, Robyn R.

    1993-01-01

    This paper analyzes the root causes of safety-related software errors in safety-critical, embedded systems. The results show that software errors identified as potentially hazardous to the system tend to be produced by different error mechanisms than non- safety-related software errors. Safety-related software errors are shown to arise most commonly from (1) discrepancies between the documented requirements specifications and the requirements needed for correct functioning of the system and (2) misunderstandings of the software's interface with the rest of the system. The paper uses these results to identify methods by which requirements errors can be prevented. The goal is to reduce safety-related software errors and to enhance the safety of complex, embedded systems.

  19. Modernization of tank floor scanning system (TAFLOSS) software

    International Nuclear Information System (INIS)

    Mohd Fitri Abdul Rahman; Jaafar Abdullah; Susan Maria Sipaun

    2002-01-01

    Tank Floor Scanning System (TAFLOSS) is a portable nucleonic device based on the scattering and moderation phenomena of neutrons. TAFLOSS, which was developed by MINT, can precisely and non-destructively measure the gap and hydrogen content in the foundation of a gigantic industrial tank in a practical and cost-effective manner. In recording and analysing measured data, three different computer software were used. In analysing the initial data, a Disk Operating System (DOS) based software called MesTank 3.0 have been developed. The system also used commercial software such as Table Curve 2D and SURFER for graphics purposes. Table Curve 2D was used to plot and evaluate curve fitting, whereas SURFER software used to draw contours. It is not user friendly and time consuming to switch from a software to another software for different tasks of this system. Therefore, the main objective of the project is to develop new user-friendly software that combined the old and commercial software into a single package. The computer programming language that was used to develop the software is Microsoft Visual C++ ver. 6.0. The process of developing this software involved complex mathematical calculation, curve fitting and contour plot. This paper describes the initial development of a computer programme for analysing the initial data and plotting exponential curve fitting. (Author)

  20. Packaging of control system software

    International Nuclear Information System (INIS)

    Zagar, K.; Kobal, M.; Saje, N.; Zagar, A.; Sabjan, R.; Di Maio, F.; Stepanov, D.

    2012-01-01

    Control system software consists of several parts - the core of the control system, drivers for integration of devices, configuration for user interfaces, alarm system, etc. Once the software is developed and configured, it must be installed to computers where it runs. Usually, it is installed on an operating system whose services it needs, and also in some cases dynamically links with the libraries it provides. Operating system can be quite complex itself - for example, a typical Linux distribution consists of several thousand packages. To manage this complexity, we have decided to rely on Red Hat Package Management system (RPM) to package control system software, and also ensure it is properly installed (i.e., that dependencies are also installed, and that scripts are run after installation if any additional actions need to be performed). As dozens of RPM packages need to be prepared, we are reducing the amount of effort and improving consistency between packages through a Maven-based infrastructure that assists in packaging (e.g., automated generation of RPM SPEC files, including automated identification of dependencies). So far, we have used it to package EPICS, Control System Studio (CSS) and several device drivers. We perform extensive testing on Red Hat Enterprise Linux 5.5, but we have also verified that packaging works on CentOS and Scientific Linux. In this article, we describe in greater detail the systematic system of packaging we are using, and its particular application for the ITER CODAC Core System. (authors)

  1. SWEPP Gamma-Ray Spectrometer System software design description

    International Nuclear Information System (INIS)

    Femec, D.A.; Killian, E.W.

    1994-08-01

    To assist in the characterization of the radiological contents of contract-handled waste containers at the Stored Waste Examination Pilot Plant (SWEPP), the SWEPP Gamma-Ray Spectrometer (SGRS) System has been developed by the Radiation Measurements and Development Unit of the Idaho National Engineering Laboratory. The SGRS system software controls turntable and detector system activities. In addition to determining the concentrations of gamma-ray-emitting radionuclides, this software also calculates attenuation-corrected isotopic mass ratios of-specific interest. This document describes the software design for the data acquisition and analysis software associated with the SGRS system

  2. SWEPP Gamma-Ray Spectrometer System software design description

    Energy Technology Data Exchange (ETDEWEB)

    Femec, D.A.; Killian, E.W.

    1994-08-01

    To assist in the characterization of the radiological contents of contract-handled waste containers at the Stored Waste Examination Pilot Plant (SWEPP), the SWEPP Gamma-Ray Spectrometer (SGRS) System has been developed by the Radiation Measurements and Development Unit of the Idaho National Engineering Laboratory. The SGRS system software controls turntable and detector system activities. In addition to determining the concentrations of gamma-ray-emitting radionuclides, this software also calculates attenuation-corrected isotopic mass ratios of-specific interest. This document describes the software design for the data acquisition and analysis software associated with the SGRS system.

  3. High Confidence Software and Systems Research Needs

    Data.gov (United States)

    Networking and Information Technology Research and Development, Executive Office of the President — This White Paper presents a survey of high confidence software and systems research needs. It has been prepared by the High Confidence Software and Systems...

  4. Managing MDO Software Development Projects

    Science.gov (United States)

    Townsend, J. C.; Salas, A. O.

    2002-01-01

    Over the past decade, the NASA Langley Research Center developed a series of 'grand challenge' applications demonstrating the use of parallel and distributed computation and multidisciplinary design optimization. All but the last of these applications were focused on the high-speed civil transport vehicle; the final application focused on reusable launch vehicles. Teams of discipline experts developed these multidisciplinary applications by integrating legacy engineering analysis codes. As teams became larger and the application development became more complex with increasing levels of fidelity and numbers of disciplines, the need for applying software engineering practices became evident. This paper briefly introduces the application projects and then describes the approaches taken in project management and software engineering for each project; lessons learned are highlighted.

  5. AVNG System Software-Attribute Verification System with Information Barriers for Mass Isotopic Measurements

    International Nuclear Information System (INIS)

    Elmont, T.H.; Langner, Diana C.; MacArthur, D.W.; Mayo, D.R.; Smith, M.K.; Modenov, A.

    2005-01-01

    This report describes the software development for the plutonium attribute verification system - AVNG. A brief synopsis of the technical solution for the measurement system is presented. The main tasks for the software development that is underway are formulated. The development tasks are shown in software structural flowcharts, measurement system state diagram and a description of the software. The current status of the AVNG software development is elucidated.

  6. Diversity requirements for safety critical software-based automation systems

    International Nuclear Information System (INIS)

    Korhonen, J.; Pulkkinen, U.; Haapanen, P.

    1998-03-01

    System vendors nowadays propose software-based systems even for the most critical safety functions in nuclear power plants. Due to the nature and mechanisms of influence of software faults new methods are needed for the safety and reliability evaluation of these systems. In the research project 'Programmable automation systems in nuclear power plants (OHA)' various safety assessment methods and tools for software based systems are developed and evaluated. This report first discusses the (common cause) failure mechanisms in software-based systems, then defines fault-tolerant system architectures to avoid common cause failures, then studies the various alternatives to apply diversity and their influence on system reliability. Finally, a method for the assessment of diversity is described. Other recently published reports in OHA-report series handles the statistical reliability assessment of software based (STUK-YTO-TR 119), usage models in reliability assessment of software-based systems (STUK-YTO-TR 128) and handling of programmable automation in plant PSA-studies (STUK-YTO-TR 129)

  7. A control system verifier using automated reasoning software

    International Nuclear Information System (INIS)

    Smith, D.E.; Seeman, S.E.

    1985-08-01

    An on-line, automated reasoning software system for verifying the actions of other software or human control systems has been developed. It was demonstrated by verifying the actions of an automated procedure generation system. The verifier uses an interactive theorem prover as its inference engine with the rules included as logical axioms. Operation of the verifier is generally transparent except when the verifier disagrees with the actions of the monitored software. Testing with an automated procedure generation system demonstrates the successful application of automated reasoning software for verification of logical actions in a diverse, redundant manner. A higher degree of confidence may be placed in the verified actions of the combined system

  8. Understanding legacy liabilities

    Energy Technology Data Exchange (ETDEWEB)

    Ossi, G.J. [Venable, LLP (United States)

    2005-08-01

    Among the most immediate issues facing operations with a workforce represented by the United Mine Workers of America (UMWA) are the so-called 'legacy liabilities'. Legacy liabilities fall under two categories: retiree health care and pension. The retiree health benefit obligations fall into two categories; statutory - those created under the Coal Industry Retiree Health Benefit Act of 1992 and contractual - the 1993 Employer Benefit Plan and the Individual Employer Plans. The pension liabilities are more straightforward; there are three different retirement plans in the NBCWA; the UMWA 1950 Pension Plan, the UMWA 1974 Pension Plan and the UMWA Cash Deferred Savings Plan of 1988.

  9. Conversion of HSPF Legacy Model to a Platform-Independent, Open-Source Language

    Science.gov (United States)

    Heaphy, R. T.; Burke, M. P.; Love, J. T.

    2015-12-01

    Since its initial development over 30 years ago, the Hydrologic Simulation Program - FORTAN (HSPF) model has been used worldwide to support water quality planning and management. In the United States, HSPF receives widespread endorsement as a regulatory tool at all levels of government and is a core component of the EPA's Better Assessment Science Integrating Point and Nonpoint Sources (BASINS) system, which was developed to support nationwide Total Maximum Daily Load (TMDL) analysis. However, the model's legacy code and data management systems have limitations in their ability to integrate with modern software, hardware, and leverage parallel computing, which have left voids in optimization, pre-, and post-processing tools. Advances in technology and our scientific understanding of environmental processes that have occurred over the last 30 years mandate that upgrades be made to HSPF to allow it to evolve and continue to be a premiere tool for water resource planners. This work aims to mitigate the challenges currently facing HSPF through two primary tasks: (1) convert code to a modern widely accepted, open-source, high-performance computing (hpc) code; and (2) convert model input and output files to modern widely accepted, open-source, data model, library, and binary file format. Python was chosen as the new language for the code conversion. It is an interpreted, object-oriented, hpc code with dynamic semantics that has become one of the most popular open-source languages. While python code execution can be slow compared to compiled, statically typed programming languages, such as C and FORTRAN, the integration of Numba (a just-in-time specializing compiler) has allowed this challenge to be overcome. For the legacy model data management conversion, HDF5 was chosen to store the model input and output. The code conversion for HSPF's hydrologic and hydraulic modules has been completed. The converted code has been tested against HSPF's suite of "test" runs and shown

  10. Assessing Resistance to Change during Shifting from Legacy to Open Web-Based Systems in the Air Transport Industry

    Science.gov (United States)

    Brewer, Denise

    2012-01-01

    The air transport industry (ATI) is a dynamic, communal, international, and intercultural environment in which the daily operations of airlines, airports, and service providers are dependent on information technology (IT). Many of the IT legacy systems are more than 30 years old, and current regulations and the globally distributed workplace have…

  11. Assessing software quality at each step of its life-cycle to enhance reliability of control systems

    International Nuclear Information System (INIS)

    Hardion, V.; Buteau, A.; Leclercq, N.; Abeille, G.; Pierre-Joseph, Z.; Le, S.

    2012-01-01

    A distributed software control system aims to enhance the upgrade ability and reliability by sharing responsibility between several components. The disadvantage is that it makes it harder to detect problems on a significant number of modules. With Kaizen in mind we have chosen to continuously invest in automation to obtain a complete overview of software quality despite the growth of legacy code. The development process has already been mastered by staging each life-cycle step thanks to a continuous integration server based on JENKINS and MAVEN. We enhanced this process, focusing on 3 objectives: Automatic Test, Static Code Analysis and Post-Mortem Supervision. Now, the build process automatically includes a test section to detect regressions, incorrect behaviour and integration incompatibility. The in-house TANGOUNIT project satisfies the difficulties of testing distributed components such as Tango Devices. In the next step, the programming code has to pass a complete code quality check-up. The SONAR quality server has been integrated in the process, to collect each static code analysis and display the hot topics on summary web pages. Finally, the integration of Google BREAKPAD in every TANGO Devices gives us essential statistics from crash reports and enables us to replay the crash scenarios at any time. We have already gained greater visibility on current developments. Some concrete results will be presented including reliability enhancement, better management of subcontracted software development, quicker adoption of coding standards by new developers and understanding of impacts when moving to a new technology. (authors)

  12. Software Build and Delivery Systems

    Energy Technology Data Exchange (ETDEWEB)

    Robey, Robert W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-07-10

    This presentation deals with the hierarchy of software build and delivery systems. One of the goals is to maximize the success rate of new users and developers when first trying your software. First impressions are important. Early successes are important. This also reduces critical documentation costs. This is a presentation focused on computer science and goes into detail about code documentation.

  13. Modeling of Some Chaotic Systems with AnyLogic Software

    Directory of Open Access Journals (Sweden)

    Biljana Zlatanovska

    2018-05-01

    Full Text Available The chaotic systems are already known in the theory of chaos. In our paper will be analyzed the following chaotic systems: Rossler, Chua and Chen systems. All of them are systems of ordinary differential equations. By mathematical software Mathematica and MatLab, their graphical representation as continuous dynamical systems is already known. By computer simulations, via examples, the systems will be analyzed using AnyLogic software. We would like to present the way how ordinary differential equations are modeling with AnyLogic software, as one of the simplest software for use.

  14. Replacement of the ISIS control system

    International Nuclear Information System (INIS)

    Mannix, R.P.; Barton, C.J.; Brownless, D.M.; Kerr, J.C.

    1992-01-01

    In operation since 1985, ISIS is the world's most powerful pulsed spallation neutron source. The decision has been taken to replace the existing ISIS control system, which has been in use for over ten years. The problems of such a project, given the legacy of processor specific hardware and software are discussed, a long with the problems associated with incorporating existing interface hardware into any new system. Present progress using commercial workstation based control software is presented with, an assessment of the benefits and pitfalls of such an approach. (author)

  15. Architecting Fault-Tolerant Software Systems

    NARCIS (Netherlands)

    Sözer, Hasan

    2009-01-01

    The increasing size and complexity of software systems makes it hard to prevent or remove all possible faults. Faults that remain in the system can eventually lead to a system failure. Fault tolerance techniques are introduced for enabling systems to recover and continue operation when they are

  16. The primary protection system software

    International Nuclear Information System (INIS)

    Tooley, P.A.

    1992-01-01

    This paper continues the detailed description of the Primary Protection System for Sizewell-B by providing an overview of design and implementation of the software, including the features of the design process which ensure that quality is delivered by the contractor. The Nuclear Electric software assessment activities are also described. The argument for the excellence of the software is made on the basis of a quality product delivered by the equipment supplier's design process, and the confirmation of this provided by the Nuclear Electric assessment process, which is as searching and complete an examination as is reasonably practicable to achieve. (author)

  17. Compiling software for a hierarchical distributed processing system

    Science.gov (United States)

    Archer, Charles J; Blocksome, Michael A; Ratterman, Joseph D; Smith, Brian E

    2013-12-31

    Compiling software for a hierarchical distributed processing system including providing to one or more compiling nodes software to be compiled, wherein at least a portion of the software to be compiled is to be executed by one or more nodes; compiling, by the compiling node, the software; maintaining, by the compiling node, any compiled software to be executed on the compiling node; selecting, by the compiling node, one or more nodes in a next tier of the hierarchy of the distributed processing system in dependence upon whether any compiled software is for the selected node or the selected node's descendents; sending to the selected node only the compiled software to be executed by the selected node or selected node's descendent.

  18. Software control and system configuration management: A systems-wide approach

    Science.gov (United States)

    Petersen, K. L.; Flores, C., Jr.

    1984-01-01

    A comprehensive software control and system configuration management process for flight-crucial digital control systems of advanced aircraft has been developed and refined to insure efficient flight system development and safe flight operations. Because of the highly complex interactions among the hardware, software, and system elements of state-of-the-art digital flight control system designs, a systems-wide approach to configuration control and management has been used. Specific procedures are implemented to govern discrepancy reporting and reconciliation, software and hardware change control, systems verification and validation testing, and formal documentation requirements. An active and knowledgeable configuration control board reviews and approves all flight system configuration modifications and revalidation tests. This flexible process has proved effective during the development and flight testing of several research aircraft and remotely piloted research vehicles with digital flight control systems that ranged from relatively simple to highly complex, integrated mechanizations.

  19. Recording the LHCb data and software dependencies

    Science.gov (United States)

    Trisovic, Ana; Couturier, Ben; Gibson, Val; Jones, Chris

    2017-10-01

    In recent years awareness of the importance of preserving the experimental data and scientific software at CERN has been rising. To support this effort, we are presenting a novel approach to structure dependencies of the LHCb data and software to make it more accessible in the long-term future. In this paper, we detail the implementation of a graph database of these dependencies. We list the implications that can be deduced from the graph mining (such as a search for the legacy software), with emphasis on data preservation. Furthermore, we introduce a methodology of recreating the LHCb data, thus supporting reproducible research and data stewardship. Finally, we describe how this information is made available to the users on a web portal that promotes data and analysis preservation and good practise with analysis documentation.

  20. MATLAB Software Versions and Licenses for the Peregrine System |

    Science.gov (United States)

    High-Performance Computing | NREL MATLAB Software Versions and Licenses for the Peregrine System MATLAB Software Versions and Licenses for the Peregrine System Learn about the MATLAB software Peregrine is R2017b. Licenses MATLAB is proprietary software. As such, users have access to a limited number

  1. Automated software configuration in the MONSOON system

    Science.gov (United States)

    Daly, Philip N.; Buchholz, Nick C.; Moore, Peter C.

    2004-09-01

    MONSOON is the next generation OUV-IR controller project being developed at NOAO. The design is flexible, emphasizing code re-use, maintainability and scalability as key factors. The software needs to support widely divergent detector systems ranging from multi-chip mosaics (for LSST, QUOTA, ODI and NEWFIRM) down to large single or multi-detector laboratory development systems. In order for this flexibility to be effective and safe, the software must be able to configure itself to the requirements of the attached detector system at startup. The basic building block of all MONSOON systems is the PAN-DHE pair which make up a single data acquisition node. In this paper we discuss the software solutions used in the automatic PAN configuration system.

  2. Legacy question

    International Nuclear Information System (INIS)

    Healy, J.W.

    1977-01-01

    The legacy question discussed refers to the definition of appropriate actions in this generation to provide a world that will allow future generations to use the earth without excessive limitations caused by our use and disposal of potentially hazardous materials

  3. Towards plug-and-play integration of archetypes into legacy electronic health record systems: the ArchiMed experience.

    Science.gov (United States)

    Duftschmid, Georg; Chaloupka, Judith; Rinner, Christoph

    2013-01-22

    The dual model approach represents a promising solution for achieving semantically interoperable standardized electronic health record (EHR) exchange. Its acceptance, however, will depend on the effort required for integrating archetypes into legacy EHR systems. We propose a corresponding approach that: (a) automatically generates entry forms in legacy EHR systems from archetypes; and (b) allows the immediate export of EHR documents that are recorded via the generated forms and stored in the EHR systems' internal format as standardized and archetype-compliant EHR extracts. As a prerequisite for applying our approach, we define a set of basic requirements for the EHR systems. We tested our approach with an EHR system called ArchiMed and were able to successfully integrate 15 archetypes from a test set of 27. For 12 archetypes, the form generation failed owing to a particular type of complex structure (multiple repeating subnodes), which was prescribed by the archetypes but not supported by ArchiMed's data model. Our experiences show that archetypes should be customized based on the planned application scenario before their integration. This would allow problematic structures to be dissolved and irrelevant optional archetype nodes to be removed. For customization of archetypes, openEHR templates or specialized archetypes may be employed. Gaps in the data types or terminological features supported by an EHR system will often not preclude integration of the relevant archetypes. More work needs to be done on the usability of the generated forms.

  4. The contribution of instrumentation and control software to system reliability

    International Nuclear Information System (INIS)

    Fryer, M.O.

    1984-01-01

    Advanced instrumentation and control systems are usually implemented using computers that monitor the instrumentation and issue commands to control elements. The control commands are based on instrument readings and software control logic. The reliability of the total system will be affected by the software design. When comparing software designs, an evaluation of how each design can contribute to the reliability of the system is desirable. Unfortunately, the science of reliability assessment of combined hardware and software systems is in its infancy. Reliability assessment of combined hardware/software systems is often based on over-simplified assumptions about software behavior. A new method of reliability assessment of combined software/hardware systems is presented. The method is based on a procedure called fault tree analysis which determines how component failures can contribute to system failure. Fault tree analysis is a well developed method for reliability assessment of hardware systems and produces quantitative estimates of failure probability based on component failure rates. It is shown how software control logic can be mapped into a fault tree that depicts both software and hardware contributions to system failure. The new method is important because it provides a way for quantitatively evaluating the reliability contribution of software designs. In many applications, this can help guide designers in producing safer and more reliable systems. An application to the nuclear power research industry is discussed

  5. System support software for TSTA [Tritium Systems Test Assembly

    International Nuclear Information System (INIS)

    Claborn, G.W.; Mann, L.W.; Nielson, C.W.

    1987-10-01

    The fact that Tritium Systems Test Assembly (TSTA) is an experimental facility makes it impossible and undesirable to try to forecast the exact software requirements. Thus the software had to be written in a manner that would allow modifications without compromising the safety requirements imposed by the handling of tritium. This suggested a multi-level approach to the software. In this approach (much like the ISO network model) each level is isolated from the level below and above by cleanly defined interfaces. For example, the subsystem support level interfaces with the subsystem hardware through the software support level. Routines in the software support level provide operations like ''OPEN VALVE'' and CLOSE VALVE'' to the subsystem level. This isolates the subsystem level from the actual hardware. This is advantageous because changes can occur in any level without the need for propagating the change to any other level. The TSTA control system consists of the hardware level, the data conversion level, the operator interface level, and the subsystem process level. These levels are described

  6. Systems and software quality the next step for industrialisation

    CERN Document Server

    Wieczorek, Martin; Bons, Heinz

    2014-01-01

    Software and systems quality is playing an increasingly important role in the growth of almost all - profit and non-profit - organisations. Quality is vital to the success of enterprises in their markets. Most small trade and repair businesses use software systems in their administration and marketing processes. Every doctor's surgery is managing its patients using software. Banking is no longer conceivable without software. Aircraft, trucks and cars use more and more software to handle their increasingly complex technical systems. Innovation, competition and cost pressure are always present i

  7. System software of the CERN proton synchrotron control system

    International Nuclear Information System (INIS)

    Carpenter, B.E.; Cailliau, R.; Cuisinier, G.; Remmer, W.

    1984-01-01

    The PS complex consists of 10 different interconnected accelerators or storage rings, mainly controlled by the same distributed system of NORD-10 and ND-100 minicomputers. After a brief outline of the hardware, this report gives a detailed description of the system software, which is based on the SINTRAN III operating system. It describes the general layout of the software, the network, CAMAC access, programming languages, program development, and microprocessor support. It concludes with reviews of performance, documentation, organization and methods, and future prospects. (orig.)

  8. A system for automatic evaluation of simulation software

    Science.gov (United States)

    Ryan, J. P.; Hodges, B. C.

    1976-01-01

    Within the field of computer software, simulation and verification are complementary processes. Simulation methods can be used to verify software by performing variable range analysis. More general verification procedures, such as those described in this paper, can be implicitly, viewed as attempts at modeling the end-product software. From software requirement methodology, each component of the verification system has some element of simulation to it. Conversely, general verification procedures can be used to analyze simulation software. A dynamic analyzer is described which can be used to obtain properly scaled variables for an analog simulation, which is first digitally simulated. In a similar way, it is thought that the other system components and indeed the whole system itself have the potential of being effectively used in a simulation environment.

  9. The Systems Biology Research Tool: evolvable open-source software

    OpenAIRE

    Wright, J; Wagner, A

    2008-01-01

    Abstract Background Research in the field of systems biology requires software for a variety of purposes. Software must be used to store, retrieve, analyze, and sometimes even to collect the data obtained from system-level (often high-throughput) experiments. Software must also be used to implement mathematical models and algorithms required for simulation and theoretical predictions on the system-level. Results We introduce a free, easy-to-use, open-source, integrated software platform calle...

  10. Towards a lessons learned system for critical software

    International Nuclear Information System (INIS)

    Andrade, J.; Ares, J.; Garcia, R.; Pazos, J.; Rodriguez, S.; Rodriguez-Paton, A.; Silva, A.

    2007-01-01

    Failure can be a major driver for the advance of any engineering discipline and Software Engineering is no exception. But failures are useful only if lessons are learned from them. In this article we aim to make a strong defence of, and set the requirements for, lessons learned systems for safety-critical software. We also present a prototype lessons learned system that includes many of the features discussed here. We emphasize that, apart from individual organizations, lessons learned systems should target industrial sectors and even the Software Engineering community. We would like to encourage the Software Engineering community to use this kind of systems as another tool in the toolbox, which complements or enhances other approaches like, for example, standards and checklists

  11. Towards a lessons learned system for critical software

    Energy Technology Data Exchange (ETDEWEB)

    Andrade, J. [University of A Coruna. Campus de Elvina, s/n. 15071, A Coruna (Spain)]. E-mail: jag@udc.es; Ares, J. [University of A Coruna. Campus de Elvina, s/n. 15071, A Coruna (Spain)]. E-mail: juanar@udc.es; Garcia, R. [University of A Coruna. Campus de Elvina, s/n. 15071, A Coruna (Spain)]. E-mail: rafael@udc.es; Pazos, J. [Technical University of Madrid. Campus de Montegancedo, s/n. 28660, Boadilla del Monte, Madrid (Spain)]. E-mail: jpazos@fi.upm.es; Rodriguez, S. [University of A Coruna. Campus de Elvina, s/n. 15071, A Coruna (Spain)]. E-mail: santi@udc.es; Rodriguez-Paton, A. [Technical University of Madrid. Campus de Montegancedo, s/n. 28660, Boadilla del Monte, Madrid (Spain)]. E-mail: arpaton@fi.upm.es; Silva, A. [Technical University of Madrid. Campus de Montegancedo, s/n. 28660, Boadilla del Monte, Madrid (Spain)]. E-mail: asilva@fi.upm.es

    2007-07-15

    Failure can be a major driver for the advance of any engineering discipline and Software Engineering is no exception. But failures are useful only if lessons are learned from them. In this article we aim to make a strong defence of, and set the requirements for, lessons learned systems for safety-critical software. We also present a prototype lessons learned system that includes many of the features discussed here. We emphasize that, apart from individual organizations, lessons learned systems should target industrial sectors and even the Software Engineering community. We would like to encourage the Software Engineering community to use this kind of systems as another tool in the toolbox, which complements or enhances other approaches like, for example, standards and checklists.

  12. The Systems Biology Research Tool: evolvable open-source software

    Directory of Open Access Journals (Sweden)

    Wright Jeremiah

    2008-06-01

    Full Text Available Abstract Background Research in the field of systems biology requires software for a variety of purposes. Software must be used to store, retrieve, analyze, and sometimes even to collect the data obtained from system-level (often high-throughput experiments. Software must also be used to implement mathematical models and algorithms required for simulation and theoretical predictions on the system-level. Results We introduce a free, easy-to-use, open-source, integrated software platform called the Systems Biology Research Tool (SBRT to facilitate the computational aspects of systems biology. The SBRT currently performs 35 methods for analyzing stoichiometric networks and 16 methods from fields such as graph theory, geometry, algebra, and combinatorics. New computational techniques can be added to the SBRT via process plug-ins, providing a high degree of evolvability and a unifying framework for software development in systems biology. Conclusion The Systems Biology Research Tool represents a technological advance for systems biology. This software can be used to make sophisticated computational techniques accessible to everyone (including those with no programming ability, to facilitate cooperation among researchers, and to expedite progress in the field of systems biology.

  13. Comparison of Overridden Medication-related Clinical Decision Support in the Intensive Care Unit between a Commercial System and a Legacy System.

    Science.gov (United States)

    Wong, Adrian; Wright, Adam; Seger, Diane L; Amato, Mary G; Fiskio, Julie M; Bates, David

    2017-08-23

    Electronic health records (EHRs) with clinical decision support (CDS) have shown to be effective at improving patient safety. Despite this, alerts delivered as part of CDS are overridden frequently, which is of concern in the critical care population as this group may have an increased risk of harm. Our organization recently transitioned from an internally-developed EHR to a commercial system. Data comparing various EHR systems, especially after transitions between EHRs, are needed to identify areas for improvement. To compare the two systems and identify areas for potential improvement with the new commercial system at a single institution. Overridden medication-related CDS alerts were included from October to December of the systems' respective years (legacy, 2011; commercial, 2015), restricted to three intensive care units. The two systems were compared with regards to CDS presentation and override rates for four types of CDS: drug-allergy, drug-drug interaction (DDI), geriatric and renal alerts. A post hoc analysis to evaluate for adverse drug events (ADEs) potentially resulting from overridden alerts was performed for 'contraindicated' DDIs via chart review. There was a significant increase in provider exposure to alerts and alert overrides in the commercial system (commercial: n=5,535; legacy: n=1,030). Rates of overrides were higher for the allergy and DDI alerts (pcommercial system. Geriatric and renal alerts were significantly different in incidence and presentation between the two systems. No ADEs were identified in an analysis of 43 overridden contraindicated DDI alerts. The vendor system had much higher rates of both alerts and overrides, although we did not find evidence of harm in a review of DDIs which were overridden. We propose recommendations for improving our current system which may be helpful to other similar institutions; improving both alert presentation and the underlying knowledge base appear important.

  14. Software System for Vocal Rendering of Printed Documents

    Directory of Open Access Journals (Sweden)

    Marian DARDALA

    2008-01-01

    Full Text Available The objective of this paper is to present a software system architecture developed to render the printed documents in a vocal form. On the other hand, in the paper are described the software solutions that exist as software components and are necessary for documents processing as well as for multimedia device controlling used by the system. The usefulness of this system is for people with visual disabilities that can access the contents of documents without that they be printed in Braille system or to exist in an audio form.

  15. Model-driven dependability assessment of software systems

    CERN Document Server

    Bernardi, Simona; Petriu, Dorina C

    2013-01-01

    In this book, the authors present cutting-edge model-driven techniques for modeling and analysis of software dependability. Most of them are based on the use of UML as software specification language. From the software system specification point of view, such techniques exploit the standard extension mechanisms of UML (i.e., UML profiling). UML profiles enable software engineers to add non-functional properties to the software model, in addition to the functional ones. The authors detail the state of the art on UML profile proposals for dependability specification and rigorously describe the t

  16. Improving Software Systems By Flow Control Analysis

    Directory of Open Access Journals (Sweden)

    Piotr Poznanski

    2012-01-01

    Full Text Available Using agile methods during the implementation of the system that meets mission critical requirements can be a real challenge. The change in the system built of dozens or even hundreds of specialized devices with embedded software requires the cooperation of a large group of engineers. This article presents a solution that supports parallel work of groups of system analysts and software developers. Deployment of formal rules to the requirements written in natural language enables using formal analysis of artifacts being a bridge between software and system requirements. Formalism and textual form of requirements allowed the automatic generation of message flow graph for the (sub system, called the “big-picture-model”. Flow diagram analysis helped to avoid a large number of defects whose repair cost in extreme cases could undermine the legitimacy of agile methods in projects of this scale. Retrospectively, a reduction of technical debt was observed. Continuous analysis of the “big picture model” improves the control of the quality parameters of the software architecture. The article also tries to explain why the commercial platform based on UML modeling language may not be sufficient in projects of this complexity.

  17. Cosimulation of embedded system using RTOS software simulator

    Science.gov (United States)

    Wang, Shihao; Duan, Zhigang; Liu, Mingye

    2003-09-01

    Embedded system design often employs co-simulation to verify system's function; one efficient verification tool of software is Instruction Set Simulator (ISS). As a full functional model of target CPU, ISS interprets instruction of embedded software step by step, which usually is time-consuming since it simulates at low-level. Hence ISS often becomes the bottleneck of co-simulation in a complicated system. In this paper, a new software verification tools, the RTOS software simulator (RSS) was presented. The mechanism of its operation was described in a full details. In RSS method, RTOS API is extended and hardware simulator driver is adopted to deal with data-exchange and synchronism between the two simulators.

  18. Software development for a switch-based data acquisition system

    Energy Technology Data Exchange (ETDEWEB)

    Booth, A. (Superconducting Super Collider Lab., Dallas, TX (United States)); Black, D.; Walsh, D. (Fermi National Accelerator Lab., Batavia, IL (United States))

    1991-12-01

    We report on the software aspects of the development of a switch-based data acquisition system at Fermilab. This paper describes how, with the goal of providing an integrated systems engineering'' environment, several powerful software tools were put in place to facilitate extensive exploration of all aspects of the design. These tools include a simulation package, graphics package and an Expert System shell which have been integrated to provide an environment which encourages the close interaction of hardware and software engineers. This paper includes a description of the simulation, user interface, embedded software, remote procedure calls, and diagnostic software which together have enabled us to provide real-time control and monitoring of a working prototype switch-based data acquisition (DAQ) system.

  19. Legacy, resource mobilisation and the olympic movement | Girginov ...

    African Journals Online (AJOL)

    There has been a growing interest in legacies of the Olympic Games focusing on external tangible outcomes, such as the number of sport competitions, participants and jobs created. Little is still known about the equally valuable internal benefits to individuals and organisational capacities of national sport systems.

  20. A Configurable, Object-Oriented, Transportation System Software Framework

    Energy Technology Data Exchange (ETDEWEB)

    KELLY,SUZANNE M.; MYRE,JOHN W.; PRICE,MARK H.; RUSSELL,ERIC D.; SCOTT,DAN W.

    2000-08-01

    The Transportation Surety Center, 6300, has been conducting continuing research into and development of information systems for the Configurable Transportation Security and Information Management System (CTSS) project, an Object-Oriented Framework approach that uses Component-Based Software Development to facilitate rapid deployment of new systems while improving software cost containment, development reliability, compatibility, and extensibility. The direction has been to develop a Fleet Management System (FMS) framework using object-oriented technology. The goal for the current development is to provide a software and hardware environment that will demonstrate and support object-oriented development commonly in the FMS Central Command Center and Vehicle domains.

  1. An Agent Based Software Approach towards Building Complex Systems

    Directory of Open Access Journals (Sweden)

    Latika Kharb

    2015-08-01

    Full Text Available Agent-oriented techniques represent an exciting new means of analyzing, designing and building complex software systems. They have the potential to significantly improve current practice in software engineering and to extend the range of applications that can feasibly be tackled. Yet, to date, there have been few serious attempts to cast agent systems as a software engineering paradigm. This paper seeks to rectify this omission. Specifically, points to be argued include:firstly, the conceptual apparatus of agent-oriented systems is well-suited to building software solutions for complex systems and secondly, agent-oriented approaches represent a genuine advance over the current state of the art for engineering complex systems. Following on from this view, the major issues raised by adopting an agentoriented approach to software engineering are highlighted and discussed in this paper.

  2. Legacy management: An old challenge with a new focus

    International Nuclear Information System (INIS)

    Gillogly, Mari; ); Sneve, Malgorzata; Smith, Graham

    2017-01-01

    The NEA Expert Group on Legacy Management (EGLM) aims to promote a practical and optimised approach for the regulatory supervision of nuclear legacy sites and installations. NEA member countries share their experiences and approaches on legacy management and have submitted case studies to the EGLM that illustrate the common challenges and approaches of many countries. The first report of the expert group will be based on these case studies and will be released in late 2017. A new, broader focus on decommissioning and legacy management issues within the NEA is expected to take shape in early 2018, carrying forward the mission to develop and promote a practical and optimised approach for the regulatory supervision of nuclear legacy sites and installations

  3. Software on the Peregrine System | High-Performance Computing | NREL

    Science.gov (United States)

    on the Peregrine System Software on the Peregrine System NREL maintains a variety of applications environment modules for use on Peregrine. Applications View list of software applications by name and research area/discipline. Libraries View list of software libraries available for linking and loading

  4. Ensuring system security through formal software evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Howell, J A; Fuyat, C [Los Alamos National Lab., NM (United States); Elvy, M [Marble Associates, Boston, MA (United States)

    1992-01-01

    With the increasing use of computer systems and networks to process safeguards information in nuclear facilities, the issue of system and data integrity is receiving worldwide attention. Among the many considerations are validation that the software performs as intended and that the information is adequately protected. Such validations are often requested of the Safeguards Systems Group of the Los Alamos National Laboratory. This paper describes our methodology for performing these software evaluations.

  5. 78 FR 47015 - Software Requirement Specifications for Digital Computer Software Used in Safety Systems of...

    Science.gov (United States)

    2013-08-02

    ... NUCLEAR REGULATORY COMMISSION [NRC-2012-0195] Software Requirement Specifications for Digital Computer Software Used in Safety Systems of Nuclear Power Plants AGENCY: Nuclear Regulatory Commission... issuing a revised regulatory guide (RG), revision 1 of RG 1.172, ``Software Requirement Specifications for...

  6. Availability of software services for a hospital information system.

    Science.gov (United States)

    Sakamoto, N

    1998-03-01

    Hospital information systems (HISs) are becoming more important and covering more parts in daily hospital operations as order-entry systems become popular and electronic charts are introduced. Thus, HISs today need to be able to provide necessary services for hospital operations for a 24-h day, 365 days a year. The provision of services discussed here does not simply mean the availability of computers, in which all that matters is that the computer is functioning. It means the provision of necessary information for hospital operations by the computer software, and we will call it the availability of software services. HISs these days are mostly client-server systems. To increase availability of software services in these systems, it is not enough to just use system structures that are highly reliable in existing host-centred systems. Four main components which support availability of software services are network systems, client computers, server computers, and application software. In this paper, we suggest how to structure these four components to provide the minimum requested software services even if a part of the system stops to function. The network system should be double-protected in stratus using Asynchronous Transfer Mode (ATM) as its base network. Client computers should be fat clients with as much application logic as possible, and reference information which do not require frequent updates (master files, for example) should be replicated in clients. It would be best if all server computers could be double-protected. However, if that is physically impossible, one database file should be made accessible by several server computers. Still, at least the basic patients' information and the latest clinical records should be double-protected physically. Application software should be tested carefully before introduction. Different versions of the application software should always be kept and managed in case the new version has problems. If a hospital

  7. Application of automated reasoning software: procedure generation system verifier

    International Nuclear Information System (INIS)

    Smith, D.E.; Seeman, S.E.

    1984-09-01

    An on-line, automated reasoning software system for verifying the actions of other software or human control systems has been developed. It was demonstrated by verifying the actions of an automated procedure generation system. The verifier uses an interactive theorem prover as its inference engine with the rules included as logic axioms. Operation of the verifier is generally transparent except when the verifier disagrees with the actions of the monitored software. Testing with an automated procedure generation system demonstrates the successful application of automated reasoning software for verification of logical actions in a diverse, redundant manner. A higher degree of confidence may be placed in the verified actions gathered by the combined system

  8. Fluidics and heat generation of Alcon Infiniti and Legacy, Bausch & Lomb Millennium, and advanced medical optics sovereign phacoemulsification systems.

    Science.gov (United States)

    Floyd, Michael S; Valentine, Jeremy R; Olson, Randall J

    2006-09-01

    To study heat generation, vacuum, and flow characteristics of the Alcon Infiniti and Bausch & Lomb Millennium with results compared with the Alcon Legacy and advanced medical optics (AMO) Sovereign machines previously studied. Experimental study. Heat generation with continuous ultrasound was determined with and without a 200-g weight. Flow and vacuum were determined from 12 to 40-ml/min in 2-ml/min steps. The impact of a STAAR Cruise Control was also tested. Millennium created the most heat/20% of power (5.67 +/- 0.51 degrees C unweighted and 6.80 +/- 0.80 degrees C weighted), followed by Sovereign (4.59 +/- 0.70 degrees C unweighted and 5.65 +/- 0.72 degrees C weighted), Infiniti (2.79 +/- 0.62 degrees C unweighted and 3.96 +/- 0.31 degrees C weighted), and Legacy (1.99 +/- 0.49 degrees C unweighted and 4.27 +/- 0.76 degrees C weighted; P Infiniti vs Legacy, both weighted). Flow studies revealed that Millennium Peristaltic was 17% less than indicated (P < .0001 to all other machines), and all other machines were within 3.5% of indicated. Cruise Control decreased flow by 4.1% (P < .0001 for same machine without it). Millennium Venturi had the greatest vacuum (81% more than the least Sovereign; P < .0001), and Cruise Control increased vacuum in a peristaltic machine 35% more than the Venturi system (P < .0001). Percent power is not consistent in regard to heat generation, however, flow was accurate for all machines except Millennium Peristaltic. Restriction with Cruise Control elevates unoccluded vacuum to levels greater than the Venturi system tested.

  9. MPS [Multiparticle Spectrometer] data acquisition software system

    International Nuclear Information System (INIS)

    Saulys, A.C.; Etkin, A.; Foley, K.J.

    1989-01-01

    A description of the software for a FASTBUS based data acquisition system in use at the Brookhaven National Laboratory Multiparticle Spectrometer is presented. Data reading and formatting is done by the SLAC Scanner Processors (SSP's) resident in the FASTBUS system. A multiprocess software system on VAX computers is used to communicate with the SSP's, record the data, and monitor on-line the progress of high energy and heavy ion experiments. The structure and the performance of this system are discussed. 4 refs., 1 fig

  10. Storage system software solutions for high-end user needs

    Science.gov (United States)

    Hogan, Carole B.

    1992-01-01

    Today's high-end storage user is one that requires rapid access to a reliable terabyte-capacity storage system running in a distributed environment. This paper discusses conventional storage system software and concludes that this software, designed for other purposes, cannot meet high-end storage requirements. The paper also reviews the philosophy and design of evolving storage system software. It concludes that this new software, designed with high-end requirements in mind, provides the potential for solving not only the storage needs of today but those of the foreseeable future as well.

  11. A hybrid approach to quantify software reliability in nuclear safety systems

    International Nuclear Information System (INIS)

    Arun Babu, P.; Senthil Kumar, C.; Murali, N.

    2012-01-01

    Highlights: ► A novel method to quantify software reliability using software verification and mutation testing in nuclear safety systems. ► Contributing factors that influence software reliability estimate. ► Approach to help regulators verify the reliability of safety critical software system during software licensing process. -- Abstract: Technological advancements have led to the use of computer based systems in safety critical applications. As computer based systems are being introduced in nuclear power plants, effective and efficient methods are needed to ensure dependability and compliance to high reliability requirements of systems important to safety. Even after several years of research, quantification of software reliability remains controversial and unresolved issue. Also, existing approaches have assumptions and limitations, which are not acceptable for safety applications. This paper proposes a theoretical approach combining software verification and mutation testing to quantify the software reliability in nuclear safety systems. The theoretical results obtained suggest that the software reliability depends on three factors: the test adequacy, the amount of software verification carried out and the reusability of verified code in the software. The proposed approach may help regulators in licensing computer based safety systems in nuclear reactors.

  12. Scale and legacy controls on catchment nutrient export regimes

    Science.gov (United States)

    Howden, N. J. K.; Burt, T.; Worrall, F.

    2017-12-01

    Nutrient dynamics in river catchments are complex: water and chemical fluxes are highly variable in low-order streams, but this variability declines as fluxes move through higher-order reaches. This poses a major challenge for process understanding as much effort is focussed on long-term monitoring of the main river channel (a high-order reach), and therefore the data available to support process understanding are predominantly derived from sites where much of the transient response of nutrient export is masked by the effect of averaging over both space and time. This may be further exacerbated at all scales by the accumulation of legacy nutrient sources in soils, aquifers and pore waters, where historical activities have led to nutrient accumulation where the catchment system is transport limited. Therefore it is of particular interest to investigate how the variability of nutrient export changes both with catchment scale (from low to high-order catchment streams) and with the presence of legacy sources, such that the context of infrequent monitoring on high-order streams can be better understood. This is not only a question of characterising nutrient export regimes per se, but also developing a more thorough understanding of how the concepts of scale and legacy may modify the statistical characteristics of observed responses across scales in both space and time. In this paper, we use synthetic data series and develop a model approach to consider how space and timescales combine with impacts of legacy sources to influence observed variability in catchment export. We find that: increasing space and timescales tend to reduce the observed variance in nutrient exports, due to an increase in travel times and greater mixing, and therefore averaging, of sources; increasing the influence of legacy sources inflates the variance, with the level of inflation dictated by the residence time of the respective sources.

  13. A user's guide to the GoldSim/BLT-MS integrated software package:a low-level radioactive waste disposal performance assessment model

    International Nuclear Information System (INIS)

    Knowlton, Robert G.; Arnold, Bill Walter; Mattie, Patrick D.

    2007-01-01

    Sandia National Laboratories (Sandia), a U.S. Department of Energy National Laboratory, has over 30 years experience in the assessment of radioactive waste disposal and at the time of this publication is providing assistance internationally in a number of areas relevant to the safety assessment of radioactive waste disposal systems. In countries with small radioactive waste programs, international technology transfer program efforts are often hampered by small budgets, schedule constraints, and a lack of experienced personnel. In an effort to surmount these difficulties, Sandia has developed a system that utilizes a combination of commercially available software codes and existing legacy codes for probabilistic safety assessment modeling that facilitates the technology transfer and maximizes limited available funding. Numerous codes developed and endorsed by the United States Nuclear Regulatory Commission (NRC) and codes developed and maintained by United States Department of Energy are generally available to foreign countries after addressing import/export control and copyright requirements. From a programmatic view, it is easier to utilize existing codes than to develop new codes. From an economic perspective, it is not possible for most countries with small radioactive waste disposal programs to maintain complex software, which meets the rigors of both domestic regulatory requirements and international peer review. Therefore, revitalization of deterministic legacy codes, as well as an adaptation of contemporary deterministic codes, provides a credible and solid computational platform for constructing probabilistic safety assessment models. This document is a reference users guide for the GoldSim/BLT-MS integrated modeling software package developed as part of a cooperative technology transfer project between Sandia National Laboratories and the Institute of Nuclear Energy Research (INER) in Taiwan for the preliminary assessment of several candidate low

  14. Verification and validation of control system software

    International Nuclear Information System (INIS)

    Munro, J.K. Jr.; Kisner, R.A.; Bhadtt, S.C.

    1991-01-01

    The following guidelines are proposed for verification and validation (V ampersand V) of nuclear power plant control system software: (a) use risk management to decide what and how much V ampersand V is needed; (b) classify each software application using a scheme that reflects what type and how much V ampersand V is needed; (c) maintain a set of reference documents with current information about each application; (d) use Program Inspection as the initial basic verification method; and (e) establish a deficiencies log for each software application. The following additional practices are strongly recommended: (a) use a computer-based configuration management system to track all aspects of development and maintenance; (b) establish reference baselines of the software, associated reference documents, and development tools at regular intervals during development; (c) use object-oriented design and programming to promote greater software reliability and reuse; (d) provide a copy of the software development environment as part of the package of deliverables; and (e) initiate an effort to use formal methods for preparation of Technical Specifications. The paper provides background information and reasons for the guidelines and recommendations. 3 figs., 3 tabs

  15. Automated Cryocooler Monitor and Control System Software

    Science.gov (United States)

    Britchcliffe, Michael J.; Conroy, Bruce L.; Anderson, Paul E.; Wilson, Ahmad

    2011-01-01

    This software is used in an automated cryogenic control system developed to monitor and control the operation of small-scale cryocoolers. The system was designed to automate the cryogenically cooled low-noise amplifier system described in "Automated Cryocooler Monitor and Control System" (NPO-47246), NASA Tech Briefs, Vol. 35, No. 5 (May 2011), page 7a. The software contains algorithms necessary to convert non-linear output voltages from the cryogenic diode-type thermometers and vacuum pressure and helium pressure sensors, to temperature and pressure units. The control function algorithms use the monitor data to control the cooler power, vacuum solenoid, vacuum pump, and electrical warm-up heaters. The control algorithms are based on a rule-based system that activates the required device based on the operating mode. The external interface is Web-based. It acts as a Web server, providing pages for monitor, control, and configuration. No client software from the external user is required.

  16. Software fault tolerance in computer operating systems

    Science.gov (United States)

    Iyer, Ravishankar K.; Lee, Inhwan

    1994-01-01

    This chapter provides data and analysis of the dependability and fault tolerance for three operating systems: the Tandem/GUARDIAN fault-tolerant system, the VAX/VMS distributed system, and the IBM/MVS system. Based on measurements from these systems, basic software error characteristics are investigated. Fault tolerance in operating systems resulting from the use of process pairs and recovery routines is evaluated. Two levels of models are developed to analyze error and recovery processes inside an operating system and interactions among multiple instances of an operating system running in a distributed environment. The measurements show that the use of process pairs in Tandem systems, which was originally intended for tolerating hardware faults, allows the system to tolerate about 70% of defects in system software that result in processor failures. The loose coupling between processors which results in the backup execution (the processor state and the sequence of events occurring) being different from the original execution is a major reason for the measured software fault tolerance. The IBM/MVS system fault tolerance almost doubles when recovery routines are provided, in comparison to the case in which no recovery routines are available. However, even when recovery routines are provided, there is almost a 50% chance of system failure when critical system jobs are involved.

  17. Software Sub-system in Loading Automatic Test System for the Measurement of Power Line Filters

    Directory of Open Access Journals (Sweden)

    Yu Bo

    2017-01-01

    Full Text Available The loading automatic test system for measurement of power line filters are in urgent demand. So the software sub-system of the whole test system was proposed. Methods: structured the test system based on the virtual instrument framework, which consisted of lower and up computer and adopted the top down approach of design to perform the system and its modules, according to the measurement principle of the test system. Results: The software sub-system including human machine interface, data analysis and process software, expert system, communication software, control software in lower computer, etc. had been designed. Furthermore, it had been integrated into the entire test system. Conclusion: This sub-system provided a fiendly software platform for the whole test system, and had many advantages such as strong functions, high performances, low prices. It not only raises the test efficiency of EMI filters, but also renders some creativities.

  18. A Reusable Software Architecture for Small Satellite AOCS Systems

    DEFF Research Database (Denmark)

    Alminde, Lars; Bendtsen, Jan Dimon; Laursen, Karl Kaas

    2006-01-01

    This paper concerns the software architecture called Sophy, which is an abbreviation for Simulation, Observation, and Planning in HYbrid systems. We present a framework that allows execution of hybrid dynamical systems in an on-line distributed computing environment, which includes interaction...... with both hardware and on-board software. Some of the key issues addressed by the framework are automatic translation of mathematical specifications of hybrid systems into executable software entities, management of execution of coupled models in a parallel distributed environment, as well as interaction...... with external components, hardware and/or software, through generic interfaces. Sophy is primarily intended as a tool for development of model based reusable software for the control and autonomous functions of satellites and/or satellite clusters....

  19. Using Software Architectures for Designing Distributed Embedded Systems

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak

    In this paper, we outline an on-going project of designing distributed embedded systems for closed-loop process control. The project is a joint effort between software architecture researchers and developers from two companies that produce commercial embedded process control systems. The project...... has a strong emphasis on software architectural issues and terminology in order to envision, design and analyze design alternatives. We present two results. First, we outline how focusing on software architecture, architectural issues and qualities are beneficial in designing distributed, embedded......, systems. Second, we present two different architectures for closed-loop process control and discuss benefits and reliabilities....

  20. Software reliability and safety in nuclear reactor protection systems

    Energy Technology Data Exchange (ETDEWEB)

    Lawrence, J.D. [Lawrence Livermore National Lab., CA (United States)

    1993-11-01

    Planning the development, use and regulation of computer systems in nuclear reactor protection systems in such a way as to enhance reliability and safety is a complex issue. This report is one of a series of reports from the Computer Safety and Reliability Group, Lawrence Livermore that investigates different aspects of computer software in reactor National Laboratory, that investigates different aspects of computer software in reactor protection systems. There are two central themes in the report, First, software considerations cannot be fully understood in isolation from computer hardware and application considerations. Second, the process of engineering reliability and safety into a computer system requires activities to be carried out throughout the software life cycle. The report discusses the many activities that can be carried out during the software life cycle to improve the safety and reliability of the resulting product. The viewpoint is primarily that of the assessor, or auditor.

  1. Software reliability and safety in nuclear reactor protection systems

    International Nuclear Information System (INIS)

    Lawrence, J.D.

    1993-11-01

    Planning the development, use and regulation of computer systems in nuclear reactor protection systems in such a way as to enhance reliability and safety is a complex issue. This report is one of a series of reports from the Computer Safety and Reliability Group, Lawrence Livermore that investigates different aspects of computer software in reactor National Laboratory, that investigates different aspects of computer software in reactor protection systems. There are two central themes in the report, First, software considerations cannot be fully understood in isolation from computer hardware and application considerations. Second, the process of engineering reliability and safety into a computer system requires activities to be carried out throughout the software life cycle. The report discusses the many activities that can be carried out during the software life cycle to improve the safety and reliability of the resulting product. The viewpoint is primarily that of the assessor, or auditor

  2. Life cycle assessment of integrated waste management systems for alternative legacy scenarios of the London Olympic Park

    Energy Technology Data Exchange (ETDEWEB)

    Parkes, Olga, E-mail: o.parkes@ucl.ac.uk; Lettieri, Paola, E-mail: p.lettieri@ucl.ac.uk; Bogle, I. David L.

    2015-06-15

    Highlights: • Application of LCA in planning integrated waste management systems. • Environmental valuation of 3 legacy scenarios for the Olympic Park. • Hot-spot analysis highlights the importance of energy and materials recovery. • Most environmental savings are achieved through materials recycling. • Sensitivity analysis shows importance of waste composition and recycling rates. - Abstract: This paper presents the results of the life cycle assessment (LCA) of 10 integrated waste management systems (IWMSs) for 3 potential post-event site design scenarios of the London Olympic Park. The aim of the LCA study is to evaluate direct and indirect emissions resulting from various treatment options of municipal solid waste (MSW) annually generated on site together with avoided emissions resulting from energy, materials and nutrients recovery. IWMSs are modelled using GaBi v6.0 Product Sustainability software and results are presented based on the CML (v.Nov-10) characterisation method. The results show that IWMSs with advanced thermal treatment (ATT) and incineration with energy recovery have the lowest Global Warming Potential (GWP) than IWMSs where landfill is the primary waste treatment process. This is due to higher direct emissions and lower avoided emissions from the landfill process compared to the emissions from the thermal treatment processes. LCA results demonstrate that significant environmental savings are achieved through substitution of virgin materials with recycled ones. The results of the sensitivity analysis carried out for IWMS 1 shows that increasing recycling rate by 5%, 10% and 15% compared to the baseline scenario can reduce GWP by 8%, 17% and 25% respectively. Sensitivity analysis also shows how changes in waste composition affect the overall result of the system. The outcomes of such assessments provide decision-makers with fundamental information regarding the environmental impacts of different waste treatment options necessary for

  3. An object-oriented framework for managing cooperating legacy databases

    NARCIS (Netherlands)

    Balsters, H; de Brock, EO

    2003-01-01

    We describe a general semantic framework for precise specification of so-called database federations. A database federation provides for tight coupling of a collection of heterogeneous legacy databases into a global integrated system. Our approach to database federation is based on the UML/OCL data

  4. Generation of Embedded Hardware/Software from SystemC

    Directory of Open Access Journals (Sweden)

    Dominique Houzet

    2006-08-01

    Full Text Available Designers increasingly rely on reusing intellectual property (IP and on raising the level of abstraction to respect system-on-chip (SoC market characteristics. However, most hardware and embedded software codes are recoded manually from system level. This recoding step often results in new coding errors that must be identified and debugged. Thus, shorter time-to-market requires automation of the system synthesis from high-level specifications. In this paper, we propose a design flow intended to reduce the SoC design cost. This design flow unifies hardware and software using a single high-level language. It integrates hardware/software (HW/SW generation tools and an automatic interface synthesis through a custom library of adapters. We have validated our interface synthesis approach on a hardware producer/consumer case study and on the design of a given software radiocommunication application.

  5. Generation of Embedded Hardware/Software from SystemC

    Directory of Open Access Journals (Sweden)

    Ouadjaout Salim

    2006-01-01

    Full Text Available Designers increasingly rely on reusing intellectual property (IP and on raising the level of abstraction to respect system-on-chip (SoC market characteristics. However, most hardware and embedded software codes are recoded manually from system level. This recoding step often results in new coding errors that must be identified and debugged. Thus, shorter time-to-market requires automation of the system synthesis from high-level specifications. In this paper, we propose a design flow intended to reduce the SoC design cost. This design flow unifies hardware and software using a single high-level language. It integrates hardware/software (HW/SW generation tools and an automatic interface synthesis through a custom library of adapters. We have validated our interface synthesis approach on a hardware producer/consumer case study and on the design of a given software radiocommunication application.

  6. Software V ampersand V methods for digital plant protection system

    International Nuclear Information System (INIS)

    Kim, Hung-Jun; Han, Jai-Bok; Chun, Chong-Son; Kim, Sung; Kim, Kern-Joong.

    1997-01-01

    Careful thought must be given to software design in the development of digital based systems that play a critical role in the successful operation of nuclear power plants. To evaluate the software verification and validation methods as well as to verify its system performance capabilities for the upgrade instrumentation and control system in the Korean future nuclear power plants, the prototype Digital Plant, Protection System (DPPS) based on the Programmable Logic Controller (PLC) has been constructed. The system design description and features are briefly presented, and the software design and software verification and validation methods are focused. 6 refs., 2 figs

  7. SIMPLIFIED CHARGED PARTICLE BEAM TRANSPORT MODELING USING COMMONLY AVAILABLE COMMERCIAL SOFTWARE

    Energy Technology Data Exchange (ETDEWEB)

    D. Douglas; K. Beard; J. Eldred; P. Evtushenko; A. Jenkins; W. Moore; L. Osborne; D. Sexton; C. Tennant

    2007-06-18

    Particle beam modeling in accelerators has been the focus of considerable effort since the 1950s. Many generations of tools have resulted from this process, each leveraging both prior experience and increases in computer power. However, continuing innovation in accelerator technology results in systems that are not well described by existing tools, so the software development process is on-going. We discuss a novel response to this situation, which was encountered when Jefferson Lab began operation of its energy-recovering linacs. These machines were not readily described with legacy soft-ware; therefore a model was built using Microsoft Excel. This interactive simulation can query data from the accelerator, use it to compute machine parameters, analyze difference orbit data, and evaluate beam properties. It can also derive new accelerator tunings and rapidly evaluate the impact of changes in machine configuration. As it is spreadsheet-based, it can be easily user-modified in response to changing requirements. Examples for the JLab IR Upgrade FEL are presented.

  8. SIMPLIFIED CHARGED PARTICLE BEAM TRANSPORT MODELING USING COMMONLY AVAILABLE COMMERCIAL SOFTWARE

    International Nuclear Information System (INIS)

    D. Douglas; K. Beard; J. Eldred; P. Evtushenko; A. Jenkins; W. Moore; L. Osborne; D. Sexton; C. Tennant

    2007-01-01

    Particle beam modeling in accelerators has been the focus of considerable effort since the 1950s. Many generations of tools have resulted from this process, each leveraging both prior experience and increases in computer power. However, continuing innovation in accelerator technology results in systems that are not well described by existing tools, so the software development process is on-going. We discuss a novel response to this situation, which was encountered when Jefferson Lab began operation of its energy-recovering linacs. These machines were not readily described with legacy soft-ware; therefore a model was built using Microsoft Excel. This interactive simulation can query data from the accelerator, use it to compute machine parameters, analyze difference orbit data, and evaluate beam properties. It can also derive new accelerator tunings and rapidly evaluate the impact of changes in machine configuration. As it is spreadsheet-based, it can be easily user-modified in response to changing requirements. Examples for the JLab IR Upgrade FEL are presented

  9. THE NEED AND KEYS FOR A NEW GENERATION NETWORK ADJUSTMENT SOFTWARE

    Directory of Open Access Journals (Sweden)

    I. Colomina

    2012-07-01

    Full Text Available Orientation and calibration of photogrammetric and remote sensing instruments is a fundamental capacity of current mapping systems and a fundamental research topic. Neither digital remote sensing acquisition systems nor direct orientation gear, like INS and GNSS technologies, made block adjustment obsolete. On the contrary, the continuous flow of new primary data acquisition systems has challenged the capacity of the legacy block adjustment systems – in general network adjustment systems – in many aspects: extensibility, genericity, portability, large data sets capacity, metadata support and many others. In this article, we concentrate on the extensibility and genericity challenges that current and future network systems shall face. For this purpose we propose a number of software design strategies with emphasis on rigorous abstract modeling that help in achieving simplicity, genericity and extensibility together with the protection of intellectual proper rights in a flexible manner. We illustrate our suggestions with the general design approach of GENA, the generic extensible network adjustment system of GeoNumerics.

  10. Adaptive intrusion data system (AIDS) software routines

    International Nuclear Information System (INIS)

    Corlis, N.E.

    1980-07-01

    An Adaptive Intrusion Data System (AIDS) was developed to collect information from intrusion alarm sensors as part of an evaluation system to improve sensor performance. AIDS is a unique digital data-compression, storage, and formatting system; it also incorporates a capability for video selection and recording for assessment of the sensors monitored by the system. The system is software reprogrammable to numerous configurations that may be used for the collection of environmental, bilevel, analog, and video data. This report describes the software routines that control the different AIDS data-collection modes, the diagnostic programs to test the operating hardware, and the data format. Sample data printouts are also included

  11. Software architecture considerations for ion source control systems

    International Nuclear Information System (INIS)

    Sinclair, J.W.

    1997-09-01

    General characteristics of distributed control system software tools are examined from the perspective of ion source control system requirements. Emphasis is placed on strategies for building extensible, distributed systems in which the ion source element is one component of a larger system. Vsystem, a commercial software tool kit from Vista Control Systems was utilized extensively in the control system upgrade of the Holifield Radioactive Ion Beam Facility. Part of the control system is described and the characteristics of Vsystem are examined and compared with those of EPICS, the Experimental Physics and Industrial Control System

  12. Evolvable Neural Software System

    Science.gov (United States)

    Curtis, Steven A.

    2009-01-01

    The Evolvable Neural Software System (ENSS) is composed of sets of Neural Basis Functions (NBFs), which can be totally autonomously created and removed according to the changing needs and requirements of the software system. The resulting structure is both hierarchical and self-similar in that a given set of NBFs may have a ruler NBF, which in turn communicates with other sets of NBFs. These sets of NBFs may function as nodes to a ruler node, which are also NBF constructs. In this manner, the synthetic neural system can exhibit the complexity, three-dimensional connectivity, and adaptability of biological neural systems. An added advantage of ENSS over a natural neural system is its ability to modify its core genetic code in response to environmental changes as reflected in needs and requirements. The neural system is fully adaptive and evolvable and is trainable before release. It continues to rewire itself while on the job. The NBF is a unique, bilevel intelligence neural system composed of a higher-level heuristic neural system (HNS) and a lower-level, autonomic neural system (ANS). Taken together, the HNS and the ANS give each NBF the complete capabilities of a biological neural system to match sensory inputs to actions. Another feature of the NBF is the Evolvable Neural Interface (ENI), which links the HNS and ANS. The ENI solves the interface problem between these two systems by actively adapting and evolving from a primitive initial state (a Neural Thread) to a complicated, operational ENI and successfully adapting to a training sequence of sensory input. This simulates the adaptation of a biological neural system in a developmental phase. Within the greater multi-NBF and multi-node ENSS, self-similar ENI s provide the basis for inter-NBF and inter-node connectivity.

  13. Philosophical Remarks on Nelson Mandela's Education Legacy

    Science.gov (United States)

    Waghid, Yusef

    2014-01-01

    In this article, I reflect on Nelson Mandela's (Madiba, the clan name of Mandela) education legacy. I argue that Madiba's education legacy is constituted by three interrelated aspects: firstly, an education for non-violence guided by deliberation, compassion and reconciliation; secondly, education as responsibility towards the Other; and thirdly,…

  14. Software challenges in extreme scale systems

    International Nuclear Information System (INIS)

    Sarkar, Vivek; Harrod, William; Snavely, Allan E

    2009-01-01

    Computer systems anticipated in the 2015 - 2020 timeframe are referred to as Extreme Scale because they will be built using massive multi-core processors with 100's of cores per chip. The largest capability Extreme Scale system is expected to deliver Exascale performance of the order of 10 18 operations per second. These systems pose new critical challenges for software in the areas of concurrency, energy efficiency and resiliency. In this paper, we discuss the implications of the concurrency and energy efficiency challenges on future software for Extreme Scale Systems. From an application viewpoint, the concurrency and energy challenges boil down to the ability to express and manage parallelism and locality by exploring a range of strong scaling and new-era weak scaling techniques. For expressing parallelism and locality, the key challenges are the ability to expose all of the intrinsic parallelism and locality in a programming model, while ensuring that this expression of parallelism and locality is portable across a range of systems. For managing parallelism and locality, the OS-related challenges include parallel scalability, spatial partitioning of OS and application functionality, direct hardware access for inter-processor communication, and asynchronous rather than interrupt-driven events, which are accompanied by runtime system challenges for scheduling, synchronization, memory management, communication, performance monitoring, and power management. We conclude by discussing the importance of software-hardware co-design in addressing the fundamental challenges for application enablement on Extreme Scale systems.

  15. The use of intelligent systems for risk management in software projects

    Directory of Open Access Journals (Sweden)

    Oksana A. Gushchina

    2017-06-01

    Full Text Available Introduction: The article identifies the main risks of a software project, examines the use of different types of intelligent systems in the risk management process for software projects, discusses the basic methods used for process estimation and forecasting in the field of software engineering, identifies currently used empty expert systems, software systems for analysis and risk management of software projects. Materials and Methods: The author describes the peculiarities of risk management in the field of software engineering with involvement of intelligent systems. The intelligent techniques allow solving the control task with expert precision without the involvement of human experts. Results: The result of this work: – identification of the key risks of a software project (tax, legal, financial and commercial risks, IT risks, personnel risks, risks related to competitors, suppliers, marketing and demand and market; – investigation of the current, applied to risk management of software system projects, artificial intelligence, particularly expert systems and software tools for evaluation of the process results; – identification of the most popular empty expert systems (Clips, G2 and Leonardo and software products of the analysis of large databases (Orange, Weka, Rattle GUI, Apache Mahout, SCaViS, RapidMiner, Databionic ESOM Tools, ELKI, KNIME, Pandas and UIMA; – consideration of the cluster, correlation, regression, factor and dispersion analysis methods for the estimation and prediction of the processes of software engineering. Discussion and Conclusions: The results show the feasibility of the application of various intelligent systems in the risk management process. The analysis of methods of evaluating risks and the tendency of their application in the modern systems of intellectual analysis can serve as a start point for creating a unified system of risk management for software projects of medium and high complexity with a

  16. A flexible software architecture for tokamak discharge control systems

    International Nuclear Information System (INIS)

    Ferron, J.R.; Penaflor, B.; Walker, M.L.; Moller, J.; Butner, D.

    1995-01-01

    The software structure of the plasma control system in use on the DIII-D tokamak experiment is described. This system implements control functions through software executing in real time on one or more digital computers. The software is organized into a hierarchy that allows new control functions needed to support the DIII-D experimental program to be added easily without affecting previously implemented functions. This also allows the software to be portable in order to create control systems for other applications. The tokamak operator uses an X-windows based interface to specify the time evolution of a tokamak discharge. The interface provides a high level view for the operator that reduces the need for detailed knowledge of the control system operation. There is provision for an asynchronous change to an alternate discharge time evolution in response to an event that is detected in real time. Quality control is enhanced through off-line testing that can make use of software-based tokamak simulators

  17. eXascale PRogramming Environment and System Software (XPRESS)

    Energy Technology Data Exchange (ETDEWEB)

    Chapman, Barbara [Univ. of Houston, TX (United States); Gabriel, Edgar [Univ. of Houston, TX (United States)

    2015-11-30

    Exascale systems, with a thousand times the compute capacity of today’s leading edge petascale computers, are expected to emerge during the next decade. Their software systems will need to facilitate the exploitation of exceptional amounts of concurrency in applications, and ensure that jobs continue to run despite the occurrence of system failures and other kinds of hard and soft errors. Adapting computations at runtime to cope with changes in the execution environment, as well as to improve power and performance characteristics, is likely to become the norm. As a result, considerable innovation is required to develop system support to meet the needs of future computing platforms. The XPRESS project aims to develop and prototype a revolutionary software system for extreme-­scale computing for both exascale and strong­scaled problems. The XPRESS collaborative research project will advance the state-­of-­the-­art in high performance computing and enable exascale computing for current and future DOE mission-­critical applications and supporting systems. The goals of the XPRESS research project are to: A. enable exascale performance capability for DOE applications, both current and future, B. develop and deliver a practical computing system software X-­stack, OpenX, for future practical DOE exascale computing systems, and C. provide programming methods and environments for effective means of expressing application and system software for portable exascale system execution.

  18. Software Defined Common Processing System (SDCPS), Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — Coherent Logix, Incorporated proposes the Software Defined Common Processing System (SDCPS) program to facilitate the development of a Software Defined Radio...

  19. Creating a career legacy map to help assure meaningful work in nursing.

    Science.gov (United States)

    Hinds, Pamela S; Britton, Dorienda R; Coleman, Lael; Engh, Eileen; Humbel, Tina Kunze; Keller, Susan; Kelly, Katherine Patterson; Menard, Johanna; Lee, Marlene A; Roberts-Turner, Renee; Walczak, Dory

    2015-01-01

    When nurses declare a professional legacy (or what they intend to be better in health care because of their efforts), they are likely to maintain a focus on achieving their legacy and to experience meaning in the process. We depict the legacy and involved steps in creating a legacy map, which is a concrete guide forward to intended career outcomes. Informed by the "meaningful work" literature, we describe a legacy map, its function, the process to create one, and the application of a legacy map to guide careers. We also describe an administrative benefit of the legacy map-the map can be used by team leaders and members to secure needed resources and opportunities to support the desired legacy of team members. Legacy mapping can be a self-use career guidance tool for nurses and other health care professionals or a tool that links the career efforts of a team member with the career support efforts of a team leader. Copyright © 2015 Elsevier Inc. All rights reserved.

  20. Re-engineering software systems in the Department of Defense using integrated computer aided software engineering tools

    OpenAIRE

    Jennings, Charles A.

    1992-01-01

    Approved for public release; distribution is unlimited The Department of Defense (DoD) is plagues with severe cost overruns and delays in developing software systems. Existing software within Dod, some developed 15-to 20 years ago, require continual maintenance and modification. Major difficulties arise with maintaining older systems due to cryptic source code and a lack of adequate documentation. To remedy this situation, the DoD, is pursuing the integrated computer aided software engi...

  1. Optimal structure of fault-tolerant software systems

    International Nuclear Information System (INIS)

    Levitin, Gregory

    2005-01-01

    This paper considers software systems consisting of fault-tolerant components. These components are built from functionally equivalent but independently developed versions characterized by different reliability and execution time. Because of hardware resource constraints, the number of versions that can run simultaneously is limited. The expected system execution time and its reliability (defined as probability of obtaining the correct output within a specified time) strictly depend on parameters of software versions and sequence of their execution. The system structure optimization problem is formulated in which one has to choose software versions for each component and find the sequence of their execution in order to achieve the greatest system reliability subject to cost constraints. The versions are to be chosen from a list of available products. Each version is characterized by its reliability, execution time and cost. The suggested optimization procedure is based on an algorithm for determining system execution time distribution that uses the moment generating function approach and on the genetic algorithm. Both N-version programming and the recovery block scheme are considered within a universal model. Illustrated example is presented

  2. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Fishler, B

    2011-03-18

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  3. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    International Nuclear Information System (INIS)

    Fishler, B.

    2011-01-01

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  4. Failure mode and effects analysis of software-based automation systems

    International Nuclear Information System (INIS)

    Haapanen, P.; Helminen, A.

    2002-08-01

    Failure mode and effects analysis (FMEA) is one of the well-known analysis methods having an established position in the traditional reliability analysis. The purpose of FMEA is to identify possible failure modes of the system components, evaluate their influences on system behaviour and propose proper countermeasures to suppress these effects. The generic nature of FMEA has enabled its wide use in various branches of industry reaching from business management to the design of spaceships. The popularity and diverse use of the analysis method has led to multiple interpretations, practices and standards presenting the same analysis method. FMEA is well understood at the systems and hardware levels, where the potential failure modes usually are known and the task is to analyse their effects on system behaviour. Nowadays, more and more system functions are realised on software level, which has aroused the urge to apply the FMEA methodology also on software based systems. Software failure modes generally are unknown - 'software modules do not fail, they only display incorrect behaviour' - and depend on dynamic behaviour of the application. These facts set special requirements on the FMEA of software based systems and make it difficult to realise. In this report the failure mode and effects analysis is studied for the use of reliability analysis of software-based systems. More precisely, the target system of FMEA is defined to be a safety-critical software-based automation application in a nuclear power plant, implemented on an industrial automation system platform. Through a literature study the report tries to clarify the intriguing questions related to the practical use of software failure mode and effects analysis. The study is a part of the research project 'Programmable Automation System Safety Integrity assessment (PASSI)', belonging to the Finnish Nuclear Safety Research Programme (FINNUS, 1999-2002). In the project various safety assessment methods and tools for

  5. Summary of the International Conference on Software and System Processes

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; O'Connor, Rory V.; Perry, Dewayne E.

    2016-01-01

    The International Conference on Software and Systems Process (ICSSP), continuing the success of Software Process Workshop (SPW), the Software Process Modeling and Simulation Workshop (ProSim) and the International Conference on Software Process (ICSP) conference series, has become the established...... premier event in the field of software and systems engineering processes. It provides a leading forum for the exchange of research outcomes and industrial best-practices in process development from software and systems disciplines. ICSSP 2016 was held in Austin, Texas, from 14-15 May 2016, co......-located with the 38th International Conference on Software Engineering (ICSE). The theme of mICSSP 2016 was studying "Process(es) in Action" by recognizing that the AS-Planned and AS-Practiced processes can be quite different in many ways including their ows, their complexity and the evolving needs of stakeholders...

  6. Software safety analysis techniques for developing safety critical software in the digital protection system of the LMR

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Jang Soo; Cheon, Se Woo; Kim, Chang Hoi; Sim, Yun Sub

    2001-02-01

    This report has described the software safety analysis techniques and the engineering guidelines for developing safety critical software to identify the state of the art in this field and to give the software safety engineer a trail map between the code and standards layer and the design methodology and documents layer. We have surveyed the management aspects of software safety activities during the software lifecycle in order to improve the safety. After identifying the conventional safety analysis techniques for systems, we have surveyed in details the software safety analysis techniques, software FMEA(Failure Mode and Effects Analysis), software HAZOP(Hazard and Operability Analysis), and software FTA(Fault Tree Analysis). We have also surveyed the state of the art in the software reliability assessment techniques. The most important results from the reliability techniques are not the specific probability numbers generated, but the insights into the risk importance of software features. To defend against potential common-mode failures, high quality, defense-in-depth, and diversity are considered to be key elements in digital I and C system design. To minimize the possibility of CMFs and thus increase the plant reliability, we have provided D-in-D and D analysis guidelines.

  7. Software safety analysis techniques for developing safety critical software in the digital protection system of the LMR

    International Nuclear Information System (INIS)

    Lee, Jang Soo; Cheon, Se Woo; Kim, Chang Hoi; Sim, Yun Sub

    2001-02-01

    This report has described the software safety analysis techniques and the engineering guidelines for developing safety critical software to identify the state of the art in this field and to give the software safety engineer a trail map between the code and standards layer and the design methodology and documents layer. We have surveyed the management aspects of software safety activities during the software lifecycle in order to improve the safety. After identifying the conventional safety analysis techniques for systems, we have surveyed in details the software safety analysis techniques, software FMEA(Failure Mode and Effects Analysis), software HAZOP(Hazard and Operability Analysis), and software FTA(Fault Tree Analysis). We have also surveyed the state of the art in the software reliability assessment techniques. The most important results from the reliability techniques are not the specific probability numbers generated, but the insights into the risk importance of software features. To defend against potential common-mode failures, high quality, defense-in-depth, and diversity are considered to be key elements in digital I and C system design. To minimize the possibility of CMFs and thus increase the plant reliability, we have provided D-in-D and D analysis guidelines

  8. Computer software design description for the integrated control and data acquisition system LDUA system

    International Nuclear Information System (INIS)

    Aftanas, B.L.

    1998-01-01

    This Computer Software Design Description (CSDD) document provides the overview of the software design for all the software that is part of the integrated control and data acquisition system of the Light Duty Utility Arm System (LDUA). It describes the major software components and how they interface. It also references the documents that contain the detailed design description of the components

  9. Systems, methods and apparatus for developing and maintaining evolving systems with software product lines

    Science.gov (United States)

    Hinchey, Michael G. (Inventor); Rash, James L. (Inventor); Pena, Joaquin (Inventor)

    2011-01-01

    Systems, methods and apparatus are provided through which an evolutionary system is managed and viewed as a software product line. In some embodiments, the core architecture is a relatively unchanging part of the system, and each version of the system is viewed as a product from the product line. Each software product is generated from the core architecture with some agent-based additions. The result may be a multi-agent system software product line.

  10. Artificial intelligence and expert systems in-flight software testing

    Science.gov (United States)

    Demasie, M. P.; Muratore, J. F.

    1991-01-01

    The authors discuss the introduction of advanced information systems technologies such as artificial intelligence, expert systems, and advanced human-computer interfaces directly into Space Shuttle software engineering. The reconfiguration automation project (RAP) was initiated to coordinate this move towards 1990s software technology. The idea behind RAP is to automate several phases of the flight software testing procedure and to introduce AI and ES into space shuttle flight software testing. In the first phase of RAP, conventional tools to automate regression testing have already been developed or acquired. There are currently three tools in use.

  11. Software engineering for the EBR-II data acquisition system conversion

    International Nuclear Information System (INIS)

    Schorzman, W.

    1988-01-01

    The purpose of this paper is to outline how EBR-II engineering approached the data acquisition system (DAS) software conversion project with the restraints of operational transparency and six weeks for final implementation and testing. Software engineering is a relatively new discipline that provides a structured philosopy for software conversion. The software life cycle is structured into six basic steps: 1) initiation, 2) requirements definition, 3) design, 4) programming, 5) testing, and 6) operations. These steps are loosely defined and can be altered to fit specific software applications. DAS software is encompassed from three sources: 1) custom software, 2) system software, and 3) in-house application software. A data flow structure is used to describe the DAS software. The categories are: 1) software used to bring signals into the central processer, 2) software that transforms the analog data to engineering units and then logs the data in the data store, and 3) software used to transport and display the data. The focus of this paper is to describe how the conversion team used a structured engineering approach and utilized the resources available to produce a quality system on time. Although successful, the conversion process provided some pit falls and stumbling blocks. Working through these obstacles enhanced our understanding and surfaced in the form of LESSONS LEARNED, which are gracefully shared in this paper

  12. Training Requirements and Information Management System. Software user guide

    Energy Technology Data Exchange (ETDEWEB)

    Cillan, T.F.; Hodgson, M.A.

    1992-05-01

    This is the software user`s guide for the Training Requirements and Information Management System. This guide defines and describes the software operating procedures as they apply to the end user of the software program. This guide is intended as a reference tool for the user who already has an indepth knowledge of the Training Requirements and Information Management System functions and data reporting requirement.

  13. 36 CFR 1194.21 - Software applications and operating systems.

    Science.gov (United States)

    2010-07-01

    ... operating systems. 1194.21 Section 1194.21 Parks, Forests, and Public Property ARCHITECTURAL AND... Standards § 1194.21 Software applications and operating systems. (a) When software is designed to run on a... shall not disrupt or disable activated features of any operating system that are identified as...

  14. Tank monitor and control system (TMACS) software configuration management plan

    International Nuclear Information System (INIS)

    GLASSCOCK, J.A.

    1999-01-01

    This Software Configuration Management Plan (SCMP) describes the methodology for control of computer software developed and supported by the Systems Development and Integration (SD and I) organization of Lockheed Martin Services, Inc. (LMSI) for the Tank Monitor and Control System (TMACS). This plan controls changes to the software and configuration files used by TMACS. The controlled software includes the Gensym software package, Gensym knowledge base files developed for TMACS, C-language programs used by TMACS, the operating system on the production machine, language compilers, and all Windows NT commands and functions which affect the operating environment. The configuration files controlled include the files downloaded to the Acromag and Westronic field instruments

  15. Use of FPGA to face electronic component obsolescence in software based safety I and C in NPPS

    International Nuclear Information System (INIS)

    Hadj, Abdellah; Bach, Julien; Esmenjaud, Claude; Daumas, Frederic; Salauen, Patrick

    2010-01-01

    In order to extend the life time of their Nuclear Power Plants (NPPs), most utilities are looking for ways to implement the renovation of their existing Instrumentation and Control (I and C) systems. When the I and C to modernize is software based, three paths can be considered: - to keep the legacy microprocessor and limit refurbishment to the associated hardware (i.e. the I/O boards, memories and the CPU board itself), - to move to another I and C platform based on another microprocessor, - to move to a non microprocessor based I and C platform. Software based I and C provide strong advantages such as flexibility and ability to implement advanced functions, however the complexity and the decreasing life time of nowadays microprocessors, mainly developed for the needs of the personal computer market, makes difficult their use and licensing for safety digital I and C systems. Solutions based on re-engineering of legacy microprocessors, or use of microprocessors dedicated to critical application need to be considered. In order to share a prospective vision of the future of I and C systems in NPPs, Electricite de France (EDF) Research and Development division and Rolls-Royce have launched a three year cooperation program on the use of the ASIC/FPGA technology in safety I and C systems. The first step of this program addresses the ability of the ASIC/FPGA technology to provide replacement solutions for former microprocessors taking as example the replacement of the Motorola MC6800 microprocessor. This paper presents the development of an IP cloning the Motorola MC6800 microprocessor, suitable for use in the refurbishment of safety I and C equipment based on this microprocessor. (authors)

  16. Russian Experience in the Regulatory Supervision of the Uranium Legacy Sites - 12441

    Energy Technology Data Exchange (ETDEWEB)

    Kiselev, M.F.; Romanov, V.V. [Federal Medical Biological Agency, Moscow (Russian Federation); Shandala, N.K.; Titov, A.V.; Kiselev, S.M.; Seregin, V.A.; Metlyaev, E.G.; Novikova, N. [Burnasyan Federal Medical Biophysical Centre, Moscow (Russian Federation); Khokhlova, E.A. [Regional Management-107 under FMBA of Russia, Krasnokamensk (Russian Federation)

    2012-07-01

    Management of the uranium legacy is accompanied with environmental impact intensity of which depends on the amount of the waste generated, the extent of that waste localization and environmental spreading. The question is: how hazardous is such impact on the environment and human health? The criterion for safety assurance is adequate regulation of the uranium legacy. Since the establishment of the uranium industry, the well done regulatory system operates in the FMBA of Russia. Such system covers inter alia, the uranium legacy. This system includes the extent laboratory network of independent control and supervision, scientific researches, regulative practices. The current Russian normative and legal basis of the regulation and its application practice has a number of problems relating to the uranium legacy, connected firstly with the environmental remediation. To improve the regulatory system, the urgent tasks are: -To introduce the existing exposure situation into the national laws and standards in compliance with the ICRP system. - To develop criteria for site remediation and return, by stages, to uncontrolled uses. The similar criteria have been developed within the Russian-Norwegian cooperation for the purpose of remediation of the sites for temporary storage of SNF and RW. - To consider possibilities and methods of optimization for the remediation strategies under development. - To separate the special category - RW resulted from uranium ore mining and dressing. The current Russian RW classification is based on the waste subdivision in terms of the specific activities. Having in mind the new RW-specific law, we receive the opportunity to separate some special category - RW originated from the uranium mining and milling. Introduction of such category can simplify significantly the situation with management of waste of uranium mining and milling processes. Such approach is implemented in many countries and approved by IAEA. The category of 'RW originated

  17. Testing digital safety system software with a testability measure based on a software fault tree

    International Nuclear Information System (INIS)

    Sohn, Se Do; Hyun Seong, Poong

    2006-01-01

    Using predeveloped software, a digital safety system is designed that meets the quality standards of a safety system. To demonstrate the quality, the design process and operating history of the product are reviewed along with configuration management practices. The application software of the safety system is developed in accordance with the planned life cycle. Testing, which is a major phase that takes a significant time in the overall life cycle, can be optimized if the testability of the software can be evaluated. The proposed testability measure of the software is based on the entropy of the importance of basic statements and the failure probability from a software fault tree. To calculate testability, a fault tree is used in the analysis of a source code. With a quantitative measure of testability, testing can be optimized. The proposed testability can also be used to demonstrate whether the test cases based on uniform partitions, such as branch coverage criteria, result in homogeneous partitions that is known to be more effective than random testing. In this paper, the testability measure is calculated for the modules of a nuclear power plant's safety software. The module testing with branch coverage criteria required fewer test cases if the module has higher testability. The result shows that the testability measure can be used to evaluate whether partitions have homogeneous characteristics

  18. Sport and exercise medicine and the Olympic health legacy

    Directory of Open Access Journals (Sweden)

    Tew Garry A

    2012-07-01

    Full Text Available Abstract London 2012 is the first Olympic and Paralympic Games to explicitly try and develop socioeconomic legacies for which success indicators are specified - the highest profile of which was to deliver a health legacy by getting two million more people more active by 2012. This editorial highlights how specialists in Sport and Exercise Medicine can contribute towards increasing physical activity participation in the UK, as well as how the National Centre for Sport and Exercise Medicine might be a useful vehicle for delivering an Olympic health legacy. Key challenges are also discussed such as acquisition of funding to support new physical activity initiatives, appropriate allocation of resources, and how to assess the impact of legacy initiatives.

  19. Development of a fatigue analysis software system

    International Nuclear Information System (INIS)

    Choi, B. I.; Lee, H. J.; Han, S. W.; Kim, J. Y.; Hwang, K. H.; Kang, J. Y.

    2001-01-01

    A general purpose fatigue analysis software to predict fatigue lives of mechanical components and structures was developed. This software has some characteristic features including functions of searching weak regions on the free surface in order to reduce computing time significantly, a database of fatigue properties for various materials, and an expert system which can assist any users to get more proper results. This software can be used in the environment consists of commercial finite element packages. Using the software developed fatigue analyses for a SAE keyhole specimen and an automobile knuckle were carried out. It was observed that the results were agree well with those from commercial packages

  20. Office of Legacy Management. Information and Records Management. Transition Guidance

    International Nuclear Information System (INIS)

    2004-01-01

    responsibilities. The DOE Office of the Chief Information Officer (OCIO) has a central role in DOE records management by providing guidance, expertise, and coordination to all DOE offices and organizations and coordination with the National Archives and Records Administration (NARA). LM and the transfer site will complete an integrated transition plan which will integrate all transition elements including information and records. As part of the overall transition plan, an Information and Records Transition Plan will be developed consistent with the integrated transition plan for the site transfer and included as an attachment. The Information and Records Management Transition Plan will be developed to assist both organizations in organizing the tasks; establishing a timetable and milestones for their completion; and identifying manpower, funding and other resources that will be needed to complete the ownership transfer. In addition, the plan will provide a valuable exchange of institutional knowledge that will assist LM in meeting the obligations of responsibly managing legacy records. Guidance for the development of the plan is included in this document. Records management concerns that may arise during site closure, such as management support, contract language and agreements, interactions with the OCIO and NARA, resource and budget considerations, and procedures to safeguard records are addressed. Guidelines and criteria for records management transition activities are also provided. These include LM expectations for the inventory, scheduling, and disposition of records; the management and transfer of electronic files, including databases and software; records finding aids, indices, and recordkeeping systems; and the process for the transfer of hard copy and electronic records to LM

  1. Office of Legacy Management. Information and Records Management. Transition Guidance

    Energy Technology Data Exchange (ETDEWEB)

    None

    2004-03-01

    information responsibilities. The DOE Office of the Chief Information Officer (OCIO) has a central role in DOE records management by providing guidance, expertise, and coordination to all DOE offices and organizations and coordination with the National Archives and Records Administration (NARA). LM and the transfer site will complete an integrated transition plan which will integrate all transition elements including information and records. As part of the overall transition plan, an Information and Records Transition Plan will be developed consistent with the integrated transition plan for the site transfer and included as an attachment. The Information and Records Management Transition Plan will be developed to assist both organizations in organizing the tasks; establishing a timetable and milestones for their completion; and identifying manpower, funding and other resources that will be needed to complete the ownership transfer. In addition, the plan will provide a valuable exchange of institutional knowledge that will assist LM in meeting the obligations of responsibly managing legacy records. Guidance for the development of the plan is included in this document. Records management concerns that may arise during site closure, such as management support, contract language and agreements, interactions with the OCIO and NARA, resource and budget considerations, and procedures to safeguard records are addressed. Guidelines and criteria for records management transition activities are also provided. These include LM expectations for the inventory, scheduling, and disposition of records; the management and transfer of electronic files, including databases and software; records finding aids, indices, and recordkeeping systems; and the process for the transfer of hard copy and electronic records to LM.

  2. Software for MR imaging system VISTA-E50

    International Nuclear Information System (INIS)

    Nakatao, Shirou; Iino, Mitsutoshi; Fukuda, Kazuhiko

    1989-01-01

    VISTA-E50 has the advantages of high-quality imaging, fast scanning, high patient throughput and easy operation featured by AI (artificial intelligence) technologies, as well as merits of compact, light-weight, space- and energy-saving system. This paper presents system software and clinical application software of VISTA-E50, especially for each function and advantage. (author)

  3. Concept of software interface for BCI systems

    Science.gov (United States)

    Svejda, Jaromir; Zak, Roman; Jasek, Roman

    2016-06-01

    Brain Computer Interface (BCI) technology is intended to control external system by brain activity. One of main part of such system is software interface, which carries about clear communication between brain and either computer or additional devices connected to computer. This paper is organized as follows. Firstly, current knowledge about human brain is briefly summarized to points out its complexity. Secondly, there is described a concept of BCI system, which is then used to build an architecture of proposed software interface. Finally, there are mentioned disadvantages of sensing technology discovered during sensing part of our research.

  4. A General Water Resources Regulation Software System in China

    Science.gov (United States)

    LEI, X.

    2017-12-01

    To avoid iterative development of core modules in water resource normal regulation and emergency regulation and improve the capability of maintenance and optimization upgrading of regulation models and business logics, a general water resources regulation software framework was developed based on the collection and analysis of common demands for water resources regulation and emergency management. It can provide a customizable, secondary developed and extensible software framework for the three-level platform "MWR-Basin-Province". Meanwhile, this general software system can realize business collaboration and information sharing of water resources regulation schemes among the three-level platforms, so as to improve the decision-making ability of national water resources regulation. There are four main modules involved in the general software system: 1) A complete set of general water resources regulation modules allows secondary developer to custom-develop water resources regulation decision-making systems; 2) A complete set of model base and model computing software released in the form of Cloud services; 3) A complete set of tools to build the concept map and model system of basin water resources regulation, as well as a model management system to calibrate and configure model parameters; 4) A database which satisfies business functions and functional requirements of general water resources regulation software can finally provide technical support for building basin or regional water resources regulation models.

  5. Software Configuration Management Plan for the Sodium Removal System

    International Nuclear Information System (INIS)

    HILL, L.F.

    2000-01-01

    This document establishers the Software Configuration Management Plan (SCMP) for the software associated with the control system of the Sodium Removal System (SRS) located in the Interim Examination and Maintenance (IEM Cell) Facility of the FFTF Flux Test

  6. Olympic Health Legacy; Essentials for Lasting Development of Host City.

    Science.gov (United States)

    Lee, Young-Hee; Kim, Jung Moon

    2013-03-01

    The purpose of the Olympic Games should be to contribute to the social development by leaving behind economic, cultural and environmental legacies to the hosting region. While tangible examples such as venues are often recognized as representative legacies of the Olympics, intangible aspects such as the environment, culture, policy and human resources have been gaining in importance. The Olympic Games, at its most fundamental level, is a sporting event. Sports not only is closely related to the physical health, but is also instrumental to fostering mental health through inspiration. One of the most important sports legacies was the general change in the population's perception on sports and physical activities; due to such change, people were able to enjoy sports as part of healthy and active everyday life and benefit physically. However, compared to tangible legacies such as the facilities, social legacies such as the general health and their planning, execution and achievements are hard to monitor. Therefore, for the Olympics to leave behind socio-cultural legacies that contribute to the development of the hosting region, there must be a thorough business plan that takes into account region-specific purpose, and is divided into stages such as before, during and after the Games. Should the 2018 Winter Olympic Games hope to create continuing contribution to its hosting region, it must leave behind 'Health Legacies' that will enhance the happiness of the hosting region's population. To this end, establishment of region-specific purpose and systematic promotion of business via detailed analysis of precedents are a must. This article aim to review the health legacy endeavors of past host cities and suggest the appropriate forms of health legacy of 2018 Pyeongchang Winter Olympic and Paralympic Games.

  7. Issues Associated with Tritium Legacy Materials

    International Nuclear Information System (INIS)

    Mills, Michael

    2008-01-01

    This paper highlights some of the issues associated with the treatment of legacy materials linked to research into tritium over many years and also of materials used to contain or store tritium. The aim of the work is to recover tritium where practicable, and to leave the residual materials passively safe, either for disposal or for continued storage. A number of materials are currently stored at AWE which either contain tritium or have been used in tritium processing. It is essential that these materials are characterised such that a strategy may be developed for their safe stewardship, and ultimately for their treatment and disposal. Treatment processes for such materials are determined by the application of best practicable means (BPM) studies in accordance with the requirements of the Environment Agency of England and Wales. Clearly, it is necessary to understand the objectives of legacy material treatment / processing and the technical options available before a definitive BPM study is implemented. The majority of tritium legacy materials with which we are concerned originate from the decommissioning of a facility that was operational from the late 1950's through to the late 1990's when, on post-operative clear-out (POCO), the entire removable and transportable tritium inventory was moved to new, purpose built facilities. One of the principle tasks to be undertaken in the new facilities is the treatment of the legacy materials to recover tritium wherever practicable, and render the residual materials passively safe for disposal or continued storage. Where tritium recovery was not reasonably or technically feasible, then a means to assure continued safe storage was to be devised and implemented. The legacy materials are in the following forms: - Uranium beds which may or may not contain adsorbed tritium gas; - Tritium gas stored in containers; - Tritide targets for neutron generation; - Tritides of a broad spectrum of metals manufactured for research / long

  8. Advanced transport operating system software upgrade: Flight management/flight controls software description

    Science.gov (United States)

    Clinedinst, Winston C.; Debure, Kelly R.; Dickson, Richard W.; Heaphy, William J.; Parks, Mark A.; Slominski, Christopher J.; Wolverton, David A.

    1988-01-01

    The Flight Management/Flight Controls (FM/FC) software for the Norden 2 (PDP-11/70M) computer installed on the NASA 737 aircraft is described. The software computes the navigation position estimates, guidance commands, those commands to be issued to the control surfaces to direct the aircraft in flight based on the modes selected on the Advanced Guidance Control System (AGSC) mode panel, and the flight path selected via the Navigation Control/Display Unit (NCDU).

  9. Prototype Software for Automated Structural Analysis of Systems

    DEFF Research Database (Denmark)

    Jørgensen, A.; Izadi-Zamanabadi, Roozbeh; Kristensen, M.

    2004-01-01

    In this paper we present a prototype software tool that is developed to analyse the structural model of automated systems in order to identify redundant information that is hence utilized for Fault detection and Isolation (FDI) purposes. The dedicated algorithms in this software tool use a tri......-partite graph that represents the structural model of the system. A component-based approach has been used to address issues such as system complexity and reconfigurability possibilities....

  10. Real time data acquisition and processing system software supported by RDOS

    International Nuclear Information System (INIS)

    Yang Huimin; Chou Gongchun; Hao Shuxiang

    1982-01-01

    The paper describes a system software supported by RDOS operating system for management and supervision a multiuser nuclear spectrum data acquisition and processing system to be made of NOVA/840 computer. The system has manifold functions including four direct address data acquisition channels. One multi-dimension nuclear data acquisition channel, four scalers and four automatic control entrances. Every channel is used competitively by each user. The system's management and all processing functions are performed by system software. The architecture of the system can renew according to users' requirement with support of the system software.This article explains the logical structure of the system software and discusses some technical keys

  11. An expert system based software sizing tool, phase 2

    Science.gov (United States)

    Friedlander, David

    1990-01-01

    A software tool was developed for predicting the size of a future computer program at an early stage in its development. The system is intended to enable a user who is not expert in Software Engineering to estimate software size in lines of source code with an accuracy similar to that of an expert, based on the program's functional specifications. The project was planned as a knowledge based system with a field prototype as the goal of Phase 2 and a commercial system planned for Phase 3. The researchers used techniques from Artificial Intelligence and knowledge from human experts and existing software from NASA's COSMIC database. They devised a classification scheme for the software specifications, and a small set of generic software components that represent complexity and apply to large classes of programs. The specifications are converted to generic components by a set of rules and the generic components are input to a nonlinear sizing function which makes the final prediction. The system developed for this project predicted code sizes from the database with a bias factor of 1.06 and a fluctuation factor of 1.77, an accuracy similar to that of human experts but without their significant optimistic bias.

  12. The art of software thermal management for embedded systems

    CERN Document Server

    Benson, Mark

    2014-01-01

    This book introduces Software Thermal Management (STM) as a means of reducing power consumption in a computing system, in order to manage heat, improve component reliability, and increase system safety.  Readers will benefit from this pragmatic guide to the field of STM for embedded systems and its catalog of software power management techniques.  Since thermal management is a key bottleneck in embedded systems design, this book focuses on power as the root cause of heat. Since software has an enormous impact on power consumption in an embedded system, this book guides readers to manage heat effectively by understanding, categorizing, and developing new ways to reduce dynamic power. Whereas most books on thermal management describe mechanisms to remove heat, this book focuses on ways to avoid generating heat in the first place.   • Explains fundamentals of software thermal management, application techniques and advanced optimization strategies; • Describes a novel method for managing dynamic power, e...

  13. Formal synthesis of application and platform behaviors of embedded software systems

    DEFF Research Database (Denmark)

    Kim, Jin Hyun; Kang, Inhye; Choi, Jin-Young

    2015-01-01

    Two main embedded software components, application software and platform software, i.e., the real-time operating system (RTOS), interact with each other in order to achieve the functionality of the system. However, they are so different in behaviors that one behavior modeling language is not suff......Two main embedded software components, application software and platform software, i.e., the real-time operating system (RTOS), interact with each other in order to achieve the functionality of the system. However, they are so different in behaviors that one behavior modeling language...... is not sufficient to model both styles of behaviors and to reason about the characteristics of their individual behaviors as well as their parallel behavior and interaction properties. In this paper, we present a formal approach to the synthesis of the application software and the RTOS behavior models...

  14. Software for managing multicrate FASTBUS Systems

    International Nuclear Information System (INIS)

    Deiss, S.R.; Gustavson, D.B.

    1982-10-01

    The FASTBUS System Manager software that was designed and implemented on an LSI-11 system using PASCAL is described. Particular attention is given to the file structures, file access mechanisms, and basic routing algorithms. Portability to other machines and languages is described

  15. Bridging the gap between legacy services and Web Services

    DEFF Research Database (Denmark)

    Bissyandé, Tegawendé; Réveillère, Laurent; Bromberg, Yérom-David

    2010-01-01

    itself. In this paper, we introduce a generative language based approach for constructing wrappers to facilitate the migration of legacy service functionalities to Web Services. To this end, we have designed the Janus domain-specific language, which provides developers with a high-level way to describe......Web Services is an increasingly used instantiation of Service-Oriented Architectures (SOA) that relies on standard Internet protocols to produce services that are highly interoperable. Other types of services, relying on legacy application layer protocols, however, cannot be composed directly....... A promising solution is to implement wrappers to translate between the application layer protocols and the WS protocol. Doing so manually, however, requires a high level of expertise, in the relevant application layer protocols, in low-level network and system programming, and in the Web Service paradigm...

  16. Analyzing Software Errors in Safety-Critical Embedded Systems

    Science.gov (United States)

    Lutz, Robyn R.

    1994-01-01

    This paper analyzes the root causes of safty-related software faults identified as potentially hazardous to the system are distributed somewhat differently over the set of possible error causes than non-safety-related software faults.

  17. Software upgrade for the DIII-D neutral beam control systems

    International Nuclear Information System (INIS)

    Cummings, J.W.; Thurgood, P.A.

    1992-01-01

    This paper reports on the Neutral Beam Software Upgrade project which was launched in early 1990. The major goals were to upgrade the MAC IV operating system to the latest revision (K.1), use standard MODCOMP software (as much as possible), and to develop a very user friendly, versatile system. Accomplishing these goals required new software to be developed and modifications to existing applications software to make it compatible with the latest operating system. The custom operating system modules to handle the message service and interrupt handling were replaced by the standard MODCOMP Inter Task Communication (ITC) and interrupt routines that are part of the MAX IV operating system. The message service provides the mechanism for doing shot task sequencing (task scheduling). The interrupt routines are used to connect external irterrupts to the system

  18. Uncoupling File System Components for Bridging Legacy and Modern Storage Architectures

    Science.gov (United States)

    Golpayegani, N.; Halem, M.; Tilmes, C.; Prathapan, S.; Earp, D. N.; Ashkar, J. S.

    2016-12-01

    Long running Earth Science projects can span decades of architectural changes in both processing and storage environments. As storage architecture designs change over decades such projects need to adjust their tools, systems, and expertise to properly integrate such new technologies with their legacy systems. Traditional file systems lack the necessary support to accommodate such hybrid storage infrastructure resulting in more complex tool development to encompass all possible storage architectures used for the project. The MODIS Adaptive Processing System (MODAPS) and the Level 1 and Atmospheres Archive and Distribution System (LAADS) is an example of a project spanning several decades which has evolved into a hybrid storage architecture. MODAPS/LAADS has developed the Lightweight Virtual File System (LVFS) which ensures a seamless integration of all the different storage architectures, including standard block based POSIX compliant storage disks, to object based architectures such as the S3 compliant HGST Active Archive System, and the Seagate Kinetic disks utilizing the Kinetic Protocol. With LVFS, all analysis and processing tools used for the project continue to function unmodified regardless of the underlying storage architecture enabling MODAPS/LAADS to easily integrate any new storage architecture without the costly need to modify existing tools to utilize such new systems. Most file systems are designed as a single application responsible for using metadata to organizing the data into a tree, determine the location for data storage, and a method of data retrieval. We will show how LVFS' unique approach of treating these components in a loosely coupled fashion enables it to merge different storage architectures into a single uniform storage system which bridges the underlying hybrid architecture.

  19. Prototype Software for Automated Structural Analysis of Systems

    DEFF Research Database (Denmark)

    Jørgensen, A.; Izadi-Zamanabadi, Roozbeh; Kristensen, M.

    2004-01-01

    In this paper we present a prototype software tool that is developed to analyse the structural model of automated systems in order to identify redundant information that is hence utilized for Fault detection and Isolation (FDI) purposes. The dedicated algorithms in this software tool use a tri......-partite graph that represents the structural model of the system. A component-based approach has been used to address issues such as system complexity and recon¯gurability possibilities....

  20. Enhancing requirements engineering for patient registry software systems with evidence-based components.

    Science.gov (United States)

    Lindoerfer, Doris; Mansmann, Ulrich

    2017-07-01

    Patient registries are instrumental for medical research. Often their structures are complex and their implementations use composite software systems to meet the wide spectrum of challenges. Commercial and open-source systems are available for registry implementation, but many research groups develop their own systems. Methodological approaches in the selection of software as well as the construction of proprietary systems are needed. We propose an evidence-based checklist, summarizing essential items for patient registry software systems (CIPROS), to accelerate the requirements engineering process. Requirements engineering activities for software systems follow traditional software requirements elicitation methods, general software requirements specification (SRS) templates, and standards. We performed a multistep procedure to develop a specific evidence-based CIPROS checklist: (1) A systematic literature review to build a comprehensive collection of technical concepts, (2) a qualitative content analysis to define a catalogue of relevant criteria, and (3) a checklist to construct a minimal appraisal standard. CIPROS is based on 64 publications and covers twelve sections with a total of 72 items. CIPROS also defines software requirements. Comparing CIPROS with traditional software requirements elicitation methods, SRS templates and standards show a broad consensus but differences in issues regarding registry-specific aspects. Using an evidence-based approach to requirements engineering for registry software adds aspects to the traditional methods and accelerates the software engineering process for registry software. The method we used to construct CIPROS serves as a potential template for creating evidence-based checklists in other fields. The CIPROS list supports developers in assessing requirements for existing systems and formulating requirements for their own systems, while strengthening the reporting of patient registry software system descriptions. It may be

  1. A Reference Model for Software and System Inspections. White Paper

    Science.gov (United States)

    He, Lulu; Shull, Forrest

    2009-01-01

    Software Quality Assurance (SQA) is an important component of the software development process. SQA processes provide assurance that the software products and processes in the project life cycle conform to their specified requirements by planning, enacting, and performing a set of activities to provide adequate confidence that quality is being built into the software. Typical techniques include: (1) Testing (2) Simulation (3) Model checking (4) Symbolic execution (5) Management reviews (6) Technical reviews (7) Inspections (8) Walk-throughs (9) Audits (10) Analysis (complexity analysis, control flow analysis, algorithmic analysis) (11) Formal method Our work over the last few years has resulted in substantial knowledge about SQA techniques, especially the areas of technical reviews and inspections. But can we apply the same QA techniques to the system development process? If yes, what kind of tailoring do we need before applying them in the system engineering context? If not, what types of QA techniques are actually used at system level? And, is there any room for improvement.) After a brief examination of the system engineering literature (especially focused on NASA and DoD guidance) we found that: (1) System and software development process interact with each other at different phases through development life cycle (2) Reviews are emphasized in both system and software development. (Figl.3). For some reviews (e.g. SRR, PDR, CDR), there are both system versions and software versions. (3) Analysis techniques are emphasized (e.g. Fault Tree Analysis, Preliminary Hazard Analysis) and some details are given about how to apply them. (4) Reviews are expected to use the outputs of the analysis techniques. In other words, these particular analyses are usually conducted in preparation for (before) reviews. The goal of our work is to explore the interaction between the Quality Assurance (QA) techniques at the system level and the software level.

  2. 77 FR 50724 - Developing Software Life Cycle Processes for Digital Computer Software Used in Safety Systems of...

    Science.gov (United States)

    2012-08-22

    ... review of applications for permits and licenses. The DG entitled ``Developing Software Life Cycle... NUCLEAR REGULATORY COMMISSION [NRC-2012-0195] Developing Software Life Cycle Processes for Digital Computer Software Used in Safety Systems of Nuclear Power Plants AGENCY: Nuclear Regulatory Commission...

  3. Software for airborne radiation monitoring system

    International Nuclear Information System (INIS)

    Sheinfeld, M.; Kadmon, Y.; Tirosh, D.; Elhanany, I.; Gabovitch, A.; Barak, D.

    1997-01-01

    The Airborne Radiation Monitoring System monitors radioactive contamination in the air or on the ground. The contamination source can be a radioactive plume or an area contaminated with radionuclides. This system is composed of two major parts: Airborne Unit carried by a helicopter, and Ground Station carried by a truck. The Airborne software is intended to be the core of a computerized airborne station. The software is written in C++ under MS-Windows with object-oriented methodology. It has been designed to be user-friendly: function keys and other accelerators are used for vital operations, a help file and help subjects are available, the Human-Machine-Interface is plain and obvious. (authors)

  4. 78 FR 47011 - Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants

    Science.gov (United States)

    2013-08-02

    ... NUCLEAR REGULATORY COMMISSION [NRC-2012-0195] Software Unit Testing for Digital Computer Software... revised regulatory guide (RG), revision 1 of RG 1.171, ``Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants.'' This RG endorses American National Standards...

  5. 77 FR 50722 - Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants

    Science.gov (United States)

    2012-08-22

    ... NUCLEAR REGULATORY COMMISSION [NRC-2012-0195] Software Unit Testing for Digital Computer Software...) is issuing for public comment draft regulatory guide (DG), DG-1208, ``Software Unit Testing for Digital Computer Software used in Safety Systems of Nuclear Power Plants.'' The DG-1208 is proposed...

  6. Architecture of a software quench management system

    International Nuclear Information System (INIS)

    Jerzy M. Nogiec et al.

    2001-01-01

    Testing superconducting accelerator magnets is inherently coupled with the proper handling of quenches; i.e., protecting the magnet and characterizing the quench process. Therefore, software implementations must include elements of both data acquisition and real-time controls. The architecture of the quench management software developed at Fermilab's Magnet Test Facility is described. This system consists of quench detection, quench protection, and quench characterization components that execute concurrently in a distributed system. Collaboration between the elements of quench detection, quench characterization and current control are discussed, together with a schema of distributed saving of various quench-related data. Solutions to synchronization and reliability in such a distributed quench system are also presented

  7. A Methodological Framework for Software Safety in Safety Critical Computer Systems

    OpenAIRE

    P. V. Srinivas Acharyulu; P. Seetharamaiah

    2012-01-01

    Software safety must deal with the principles of safety management, safety engineering and software engineering for developing safety-critical computer systems, with the target of making the system safe, risk-free and fail-safe in addition to provide a clarified differentaition for assessing and evaluating the risk, with the principles of software risk management. Problem statement: Prevailing software quality models, standards were not subsisting in adequately addressing the software safety ...

  8. Achieving Critical System Survivability Through Software Architectures

    National Research Council Canada - National Science Library

    Knight, John C; Strunk, Elisabeth A

    2006-01-01

    .... In a system with a survivability architecture, under adverse conditions such as system damage or software failures, some desirable function will be eliminated but critical services will be retained...

  9. Guidelines for the verification and validation of expert system software and conventional software. Volume 7, User's manual: Final report

    International Nuclear Information System (INIS)

    Miller, L.A.; Hayes, J.E.; Mirsky, S.M.

    1995-05-01

    Reliable software is required for nuclear power industry applications. Verification and validation techniques applied during the software development process can help eliminate errors that could inhibit the proper operation of digital systems and cause availability and safety problems. Most of the techniques described in this report are valid for conventional software systems as well as for expert systems. The project resulted in a set of 16 V ampersand V guideline packages and 11 sets of procedures based on the class, development phase, and system component being tested. These guideline packages and procedures help a utility define the level of V ampersand V, which involves evaluating the complexity and type of software component along with the consequences of failure. In all, the project identified 153 V ampersand V techniques for conventional software systems and demonstrated their application to all aspects of expert systems except for the knowledge base, which requires specially developed tools. Each of these conventional techniques covers anywhere from 2-52 total types of conventional software defects, and each defect is covered by 21-50 V ampersand V techniques. The project also identified automated tools to Support V ampersand V activities

  10. Analysis of Architecture Pattern Usage in Legacy System Architecture Documentation

    NARCIS (Netherlands)

    Harrison, Neil B.; Avgeriou, Paris

    2008-01-01

    Architecture patterns are an important tool in architectural design. However, while many architecture patterns have been identified, there is little in-depth understanding of their actual use in software architectures. For instance, there is no overview of how many patterns are used per system or

  11. Software for graphic display systems

    International Nuclear Information System (INIS)

    Karlov, A.A.

    1978-01-01

    In this paper some aspects of graphic display systems are discussed. The design of a display subroutine library is described, with an example, and graphic dialogue software is considered primarily from the point of view of the programmer who uses a high-level language. (Auth.)

  12. A communication-channel-based representation system for software

    NARCIS (Netherlands)

    Demirezen, Zekai; Tanik, Murat M.; Aksit, Mehmet; Skjellum, Anthony

    We observed that before initiating software development the objectives are minimally organized and developers introduce comparatively higher organization throughout the design process. To be able to formally capture this observation, a new communication channel representation system for software is

  13. After the Fall: The RHESSI Legacy Archive

    Science.gov (United States)

    Schwartz, Richard A.; Zarro, Dominic M.; Tolbert, Anne K.

    2017-08-01

    Launched in 2002 the Ramaty High Energy Solar Spectroscopic Imager (RHESSI) continues to observe the Sun with a nearly 50% duty cycle. During that time the instrument has recorded ~100,000 solar flares in energies from 4 keV to over 10 MeV.with durations of 10s to 1000s of seconds. However, for the reasons of the decline of the solar cycle, possible failure of the instrument, or the absence of funding, our operational phase will end someday. We describe here our plans to continue to serve this dataset in raw, processed, and analyzed forms to the worldwide solar community to continue our legacy of a stream of rich scientific results.We have and are providing a stream of quicklook lightcurves, spectra, and images that we mainly serve through a web interface as well as the data in raw form to be fully analyzed within our own branch of Solar Software written in IDL. We are in the process of creating higher quality images for flares in multiple energy bands on relevant timescales for those whose needs can be met without further processing. For users with IDL licenses we expect this software to be available far into the unknowable future. Together with a database of AIA cutouts during all SDO-era flares, along with software to recover saturated images by using the AIA diffraction fringes, these will be a highly used resource.We also are developing additional tools and databases that will increase the utility of RHESSI data to members of the community with and without either IDL licenses or full access to the RHESSI database. We will provide a database of RHESSI x-ray visibilities obtained during flares at a >4 second cadence over a broad range of detectable energies. With our IDL software those can be rendered as images for times and energies of nearly the analysts's choosing. And going beyond that we are converting our imaging procedures to the Python language to eliminate the need for an IDL license. We are also developing methods to allow the customization of these

  14. Lessons learned from development and quality assurance of software systems at the Halden Project

    International Nuclear Information System (INIS)

    Bjorlo, T.J.; Berg, O.; Pehrsen, M.; Dahll, G.; Sivertsen, T.

    1996-01-01

    The OECD Halden Reactor Project has developed a number of software systems within the research programmes. These programmes have comprised a wide range of topics, like studies of software for safety-critical applications, development of different operator support systems, and software systems for building and implementing graphical user interfaces. The systems have ranged from simple prototypes to installations in process plants. In the development of these software systems, Halden has gained much experience in quality assurance of different types of software. This paper summarises the accumulated experience at the Halden Project in quality assurance of software systems. The different software systems being developed at the Halden Project may be grouped into three categories. These are plant-specific software systems (one-of-a-kind deliveries), generic software products, and safety-critical software systems. This classification has been found convenient as the categories have different requirements to the quality assurance process. In addition, the experience from use of software development tools and proprietary software systems at Halden, is addressed. The paper also focuses on the experience gained from the complete software life cycle, starting with the software planning phase and ending with software operation and maintenance

  15. Overview of radium legacies in Belgium - 59367

    International Nuclear Information System (INIS)

    Dehandschutter, B.; Jadoul, L.; Mannaerts, K.; Pepin, S.; Poffijn, A.; Blommaert, W.; Sonck, M.

    2012-01-01

    The Belgian metallurgical company, Union Miniere, has been a key-player in the sector of radium production between 1922 and 1969. The factory based in Olen has extracted radium from minerals and produced radium sources during that period. The radium production facilities have been dismantled in the 70's but legacies of the former production have still to be remediated. An overview of these legacies and of their radiological characteristics will be given. Next to the sites related to radium production, other radium legacies are related to NORM industries, essentially from the phosphate sector (phosphogypsum and CaF 2 stacks). The issue of radium legacies in Belgium encompasses a variety of concrete situations. Next to the issue of the legacies of the former radium production, the other radium contaminated sites are related to current or former NORM industries, especially from the phosphate sector. The methodological and regulatory approaches towards these sites have been described elsewhere in these proceedings. The outcome differs according to the specificities of the site: it will not be the same for the legacies of former radium production where the inventory of radioactivity includes materials which have to be considered and treated as radioactive waste (for example, disused radium sources) than for phosphogypsum stacks where a sufficient level of protection may be brought by relatively simple measures such as restrictions on the use of the site. For these sites, like PG stacks, where radon is the most important exposure pathway in case of intrusion scenario, regulatory measures similar to the ones applied to 'radon-prone areas' (restrictions in the construction of buildings, compulsory radon monitoring in workplaces present on the site,...) may be implemented. In all cases, the radiological risk-assessment will be crucial for the decision-making process. The examples given showed that the probability of occurrence of 'intrusion scenario' (like construction of

  16. Conceptual design for controller software of mechatronic systems

    NARCIS (Netherlands)

    Broenink, Johannes F.; Hilderink, G.H.; Bakkers, André; Bradshaw, Alan; Counsell, John

    1998-01-01

    The method and software tool presented here, aims at supporting the development of control software for mechatronic systems. Heterogeneous distributed embedded processors are considered as target hardware. Principles of the method are that the implementation process is a stepwise refinement from

  17. Advanced information processing system: Input/output network management software

    Science.gov (United States)

    Nagle, Gail; Alger, Linda; Kemp, Alexander

    1988-01-01

    The purpose of this document is to provide the software requirements and specifications for the Input/Output Network Management Services for the Advanced Information Processing System. This introduction and overview section is provided to briefly outline the overall architecture and software requirements of the AIPS system before discussing the details of the design requirements and specifications of the AIPS I/O Network Management software. A brief overview of the AIPS architecture followed by a more detailed description of the network architecture.

  18. Fabrication of a Sludge-Conditioning System for processing legacy wastes from the Gunite and Associated Tanks

    International Nuclear Information System (INIS)

    Randolph, J.D.; Lewis, B.E.; Farmer, J.R.; Johnson, M.A.

    2000-01-01

    The Sludge Conditioning System (SCS) for the Gunite and Associated Tanks (GAATs) is designed to receive, monitor, characterize and process legacy waste materials from the South Tank Farm tanks in preparation for final transfer of the wastes to the Melton Valley Storage Tanks (MVSTs), which are located at Oak Ridge National Laboratory. The SCS includes (1) a Primary Conditioning System (PCS) Enclosure for sampling and particle size classification, (2) a Solids Monitoring Test Loop (SMTL) for slurry characterization, (3) a Waste Transfer Pump to retrieve and transfer waste materials from GAAT consolidation tank W-9 to the MVSTs, (4) a PulsAir Mixing System to provide mixing of consolidated sludges for ease of retrieval, and (5) the interconnecting piping and valving. This report presents the design, fabrication, cost, and fabrication schedule information for the SCS

  19. U.S. Department of Energy, Office of Legacy Management Program Update, April-June 2009

    Energy Technology Data Exchange (ETDEWEB)

    None

    2009-04-01

    Welcome to the April-June 2009 issue of the U.S. Department of Energy (DOE) Office of Legacy Management (LM) Program Update. This publication is designed to provide a status of activities within LM. The Legacy Management goals are: (1) Protect human health and the environment through effective and efficient long-term surveillance and maintenance - This goal highlights DOE's responsibility to ensure long-term protection of people, the environment, and the integrity of engineered remedies and monitoring systems. (2) Preserve, protect, and make accessible legacy records and information - This goal recognizes LM's commitment to successfully manage records, information, and archives of legacy sites under its authority. (3) Support an effective and efficient work force structured to accomplish Departmental missions and assure continuity of contractor worker pension and medical benefits - This goal recognizes DOE's commitment to its contracted work force and the consistent management of pension and health benefits. As sites continue to close, DOE faces the challenges of managing pension plan and health benefits liability. (4) Manage legacy land and assets, emphasizing protective real and personal property reuse and disposition - This goal recognizes a DOE need for local collaborative management of legacy assets, including coordinating land use planning, personal property disposition to community reuse organizations, and protecting heritage resources (natural, cultural, and historical). (5) Improve program effectiveness through sound management - This goal recognizes that LM's goals cannot be attained efficiently unless the federal and contractor work force is motivated to meet requirements and work toward continuous performance improvement.

  20. Modular system for the control of complex accelerators using portable software

    International Nuclear Information System (INIS)

    von der Schmitt, H.; Aufhaus, H.

    1982-01-01

    When designing the Mainz Microtron control system, care was taken to achieve an expandable system with long-lived application software. A multi-processor system was built from the beginning. The software is split into modules, according to function and position in hierarchy, which are distributed over the computers. The decoupling which results from modularity eases software development and maintainance. RATFOR was chosen as implementation language. With a message system for communication between the modules, several aims were reached at once: (1) symbolic addressing of the accelerator components throughout the software layers, (2) transparent access to I/O devices (CAMAC) at remote computers, (3) multitasking in FORTRAN (and RATFOR) programs, (4) a separating layer for adaptation to different operating systems - essential points for software portability. The system is in operation since April 1979 for the control of MAMI stage I

  1. User and system considerations for the TCSTEK software library

    International Nuclear Information System (INIS)

    Gray, W.H.

    1979-08-01

    This report documents the idiosyncrasies of the Tektronix PLOT 10 Terminal Control System level 3.3 software as it currently exists on the ORNL Fusion Energy Division DECsystem-10 computer. It is intended to serve as a reference for future Terminal Control System updates in order that continuity between releases of Terminal Control System PLOT 10 software may be maintained

  2. Digital image processing software system using an array processor

    International Nuclear Information System (INIS)

    Sherwood, R.J.; Portnoff, M.R.; Journeay, C.H.; Twogood, R.E.

    1981-01-01

    A versatile array processor-based system for general-purpose image processing was developed. At the heart of this system is an extensive, flexible software package that incorporates the array processor for effective interactive image processing. The software system is described in detail, and its application to a diverse set of applications at LLNL is briefly discussed. 4 figures, 1 table

  3. User and system considerations for the TCSTEK software library

    Energy Technology Data Exchange (ETDEWEB)

    Gray, W.H.

    1979-08-01

    This report documents the idiosyncrasies of the Tektronix PLOT 10 Terminal Control System level 3.3 software as it currently exists on the ORNL Fusion Energy Division DECsystem-10 computer. It is intended to serve as a reference for future Terminal Control System updates in order that continuity between releases of Terminal Control System PLOT 10 software may be maintained.

  4. System and method for deriving a process-based specification

    Science.gov (United States)

    Hinchey, Michael Gerard (Inventor); Rash, James Larry (Inventor); Rouff, Christopher A. (Inventor)

    2009-01-01

    A system and method for deriving a process-based specification for a system is disclosed. The process-based specification is mathematically inferred from a trace-based specification. The trace-based specification is derived from a non-empty set of traces or natural language scenarios. The process-based specification is mathematically equivalent to the trace-based specification. Code is generated, if applicable, from the process-based specification. A process, or phases of a process, using the features disclosed can be reversed and repeated to allow for an interactive development and modification of legacy systems. The process is applicable to any class of system, including, but not limited to, biological and physical systems, electrical and electro-mechanical systems in addition to software, hardware and hybrid hardware-software systems.

  5. Advanced Transport Operating System (ATOPS) color displays software description microprocessor system

    Science.gov (United States)

    Slominski, Christopher J.; Plyler, Valerie E.; Dickson, Richard W.

    1992-01-01

    This document describes the software created for the Sperry Microprocessor Color Display System used for the Advanced Transport Operating Systems (ATOPS) project on the Transport Systems Research Vehicle (TSRV). The software delivery known as the 'baseline display system', is the one described in this document. Throughout this publication, module descriptions are presented in a standardized format which contains module purpose, calling sequence, detailed description, and global references. The global reference section includes procedures and common variables referenced by a particular module. The system described supports the Research Flight Deck (RFD) of the TSRV. The RFD contains eight cathode ray tubes (CRTs) which depict a Primary Flight Display, Navigation Display, System Warning Display, Takeoff Performance Monitoring System Display, and Engine Display.

  6. The achievement and assessment of safety in systems containing software

    International Nuclear Information System (INIS)

    Ball, A.; Dale, C.J.; Butterfield, M.H.

    1986-01-01

    In order to establish confidence in the safe operation of a reactor protection system, there is a need to establish, as far as it is possible, that: (i) the algorithms used are correct; (ii) the system is a correct implementation of the algorithms; and (iii) the hardware is sufficiently reliable. This paper concentrates principally on the second of these, as it applies to the software aspect of the more accurate and complex trip functions to be performed by modern reactor protection systems. In order to engineer safety into software, there is a need to use a development strategy which will stand a high chance of achieving a correct implementation of the trip algorithms. This paper describes three broad methodologies by which it is possible to enhance the integrity of software: fault avoidance, fault tolerance and fault removal. Fault avoidance is concerned with making the software as fault free as possible by appropriate choice of specification, design and implementation methods. A fault tolerant strategy may be advisable in many safety critical applications, in order to guard against residual faults present in the software of the installed system. Fault detection and removal techniques are used to remove as many faults as possible of those introduced during software development. The paper also discusses safety and reliability assessment as it applies to software, outlining the various approaches available. Finally, there is an outline of a research project underway in the UKAEA which is intended to assess methods for developing and testing safety and protection systems involving software. (author)

  7. Space Flight Software Development Software for Intelligent System Health Management

    Science.gov (United States)

    Trevino, Luis C.; Crumbley, Tim

    2004-01-01

    The slide presentation examines the Marshall Space Flight Center Flight Software Branch, including software development projects, mission critical space flight software development, software technical insight, advanced software development technologies, and continuous improvement in the software development processes and methods.

  8. Customizable software architectures in the accelerator control system environment

    CERN Document Server

    Mejuev, I; Kadokura, E

    2001-01-01

    Tailoring is further evolution of an application after deployment in order to adapt it to requirements that were not accounted for in the original design. End-user customization has been extensively researched in applied computer science from HCI and software engineering perspectives. Customization allows coping with flexibility requirements, decreasing maintenance and development costs of software products. In general, dynamic or diverse software requirements constitute the need for implementing end-user customization in computer systems. In accelerator physics research the factor of dynamic requirements is especially important, due to frequent software and hardware modifications resulting in correspondingly high upgrade and maintenance costs. We introduce the results of feasibility study on implementing end-user tailorability in the software for accelerator control system, considering the design and implementation of a distributed monitoring application for the 12 GeV KEK Proton Synchrotron as an example. T...

  9. Software Tools for Development on the Peregrine System | High-Performance

    Science.gov (United States)

    Computing | NREL Software Tools for Development on the Peregrine System Software Tools for and manage software at the source code level. Cross-Platform Make and SCons The "Cross-Platform Make" (CMake) package is from Kitware, and SCons is a modern software build tool based on Python

  10. Research and Development of Statistical Analysis Software System of Maize Seedling Experiment

    OpenAIRE

    Hui Cao

    2014-01-01

    In this study, software engineer measures were used to develop a set of software system for maize seedling experiments statistics and analysis works. During development works, B/S structure software design method was used and a set of statistics indicators for maize seedling evaluation were established. The experiments results indicated that this set of software system could finish quality statistics and analysis for maize seedling very well. The development of this software system explored a...

  11. Clinical software for MR imaging system, 4

    International Nuclear Information System (INIS)

    Shimizu, Koji; Kasai, Akira; Okamura, Shoichi

    1992-01-01

    Magnetic resonance imaging continues to elicit new application software through the recent technological advances of MR equipment. This paper describes several applications of our newly developed clinical software. The fast SE sequence (RISE) has proved to reduce routine examination time and to improve image quality, and ultra-fast FE sequence (SMASH) was found to extend the diagnostic capabilities in the field of cardiac study. Diffusion/perfusion imaging achieved in our MR system showed significant promise for providing novel information regarding tissue characterization. Furthermore, Image quality and practicalities of MR angiography have been improved by advanced imaging sequences and sophisticated post-processing software. (author)

  12. Software upgrade for the DIII-D neutral beam control systems

    International Nuclear Information System (INIS)

    Cummings, J.W.; Thurgood, P.A.

    1991-11-01

    The neutral beams are used to heat the plasma in the DIII-D tokamak, a fusion energy research experiment operated by General Atomics (GA) and funded by the Department of Energy (DOE). The experiment is dedicated to demonstrating noninductive current drive of high beta high temperature divertor plasma with good confinement. The neutral beam heating system for the DIII-D tokamak uses four MODCOMP Classic computers for data acquisition and control of the four beamlines. The Neutral Beam Software Upgrade project was launched in early 1990. The major goals were to upgrade the MAX IV operating system to the latest revision (K.1), use standard MODCOMP software (as much as possible), and to develop a very ''user friendly,'' versatile system. Accomplishing these goals required new software to be developed and modifications to existing applications software to make it compatible with the latest operating system. The custom operating system modules to handle the message service and interrupt handling were replaced by the standard MODCOMP Inter Task Communication (ITC) and interrupt routines that are part of the MAX IV operating system. The message service provides the mechanism for doing shot task sequencing (task scheduling). The interrupt routines are used to connect external interrupts to the system. The new software developed consists of a task dispatcher, screen manager, and interrupt tasks. The existing applications software had to be modified to be compatible with the MODCOMP ITC services and consists of the Modcomp Infinity Data Base Manager, a multi-user system, and menu-driven operating system interface routines using the Infinity Data Base Manager

  13. Olympic Health Legacy; Essentials for Lasting Development of Host City

    Science.gov (United States)

    Lee, Young-Hee; Kim, Jung Moon

    2013-01-01

    The purpose of the Olympic Games should be to contribute to the social development by leaving behind economic, cultural and environmental legacies to the hosting region. While tangible examples such as venues are often recognized as representative legacies of the Olympics, intangible aspects such as the environment, culture, policy and human resources have been gaining in importance. The Olympic Games, at its most fundamental level, is a sporting event. Sports not only is closely related to the physical health, but is also instrumental to fostering mental health through inspiration. One of the most important sports legacies was the general change in the population’s perception on sports and physical activities; due to such change, people were able to enjoy sports as part of healthy and active everyday life and benefit physically. However, compared to tangible legacies such as the facilities, social legacies such as the general health and their planning, execution and achievements are hard to monitor. Therefore, for the Olympics to leave behind socio-cultural legacies that contribute to the development of the hosting region, there must be a thorough business plan that takes into account region-specific purpose, and is divided into stages such as before, during and after the Games. Should the 2018 Winter Olympic Games hope to create continuing contribution to its hosting region, it must leave behind ‘Health Legacies’ that will enhance the happiness of the hosting region’s population. To this end, establishment of region-specific purpose and systematic promotion of business via detailed analysis of precedents are a must. This article aim to review the health legacy endeavors of past host cities and suggest the appropriate forms of health legacy of 2018 Pyeongchang Winter Olympic and Paralympic Games. PMID:26064832

  14. Testing methodology of embedded software in digital plant protection system

    International Nuclear Information System (INIS)

    Seong, Ah Young; Choi, Bong Joo; Lee, Na Young; Hwang, Il Soon

    2001-01-01

    It is necessary to assure the reliability of software in order to digitalize RPS(Reactor Protection System). Since RPS causes fatal damage on accidental cases, it is classified as Safety 1E class. Therefore we propose the effective testing methodology to assure the reliability of embedded software in the DPPS(Digital Plant Protection System). To test the embedded software effectively in DPPS, our methodology consists of two steps. The first is the re-engineering step that extracts classes from structural source program, and the second is the level of testing step which is composed of unit testing, Integration Testing and System Testing. On each testing step we test the embedded software with selected test cases after the test item identification step. If we use this testing methodology, we can test the embedded software effectively by reducing the cost and the time

  15. A dynamic system for ATLAS software installation on OSG grid sites

    International Nuclear Information System (INIS)

    Zhao, X; Maeno, T; Wenaus, T; Leuhring, F; Youssef, S; Brunelle, J; De Salvo, A; Thompson, A S

    2010-01-01

    A dynamic and reliable system for installing the ATLAS software releases on Grid sites is crucial to guarantee the timely and smooth start of ATLAS production and reduce its failure rate. In this paper, we discuss the issues encountered in the previous software installation system, and introduce the new approach, which is built upon the new development in the areas of the ATLAS workload management system (PanDA), and software package management system (pacman). It is also designed to integrate with the EGEE ATLAS software installation framework. In the new system, ATLAS software releases are packaged as pacball, a uniquely identifiable and reproducible self-installing data file. The distribution of pacballs to remote sites is managed by ATLAS data management system (DQ2) and PanDA server. The installation on remote sites is automatically triggered by the PanDA pilot jobs. The installation job payload connects to a central ATLAS software installation portal, making the information of installation status easily accessible across OSG and EGEE Grids. The issues encountered in running the new system in production, and our future plan for improvement, will also be discussed.

  16. Tank monitor and control system (TMACS) software configuration management plan; TOPICAL

    International Nuclear Information System (INIS)

    GLASSCOCK, J.A.

    1999-01-01

    This Software Configuration Management Plan (SCMP) describes the methodology for control of computer software developed and supported by the Systems Development and Integration (SD and I) organization of Lockheed Martin Services, Inc. (LMSI) for the Tank Monitor and Control System (TMACS). This plan controls changes to the software and configuration files used by TMACS. The controlled software includes the Gensym software package, Gensym knowledge base files developed for TMACS, C-language programs used by TMACS, the operating system on the production machine, language compilers, and all Windows NT commands and functions which affect the operating environment. The configuration files controlled include the files downloaded to the Acromag and Westronic field instruments

  17. Project W-211, initial tank retrieval systems, retrieval control system software configuration management plan

    International Nuclear Information System (INIS)

    RIECK, C.A.

    1999-01-01

    This Software Configuration Management Plan (SCMP) provides the instructions for change control of the W-211 Project, Retrieval Control System (RCS) software after initial approval/release but prior to the transfer of custody to the waste tank operations contractor. This plan applies to the W-211 system software developed by the project, consisting of the computer human-machine interface (HMI) and programmable logic controller (PLC) software source and executable code, for production use by the waste tank operations contractor. The plan encompasses that portion of the W-211 RCS software represented on project-specific AUTOCAD drawings that are released as part of the C1 definitive design package (these drawings are identified on the drawing list associated with each C-1 package), and the associated software code. Implementation of the plan is required for formal acceptance testing and production release. The software configuration management plan does not apply to reports and data generated by the software except where specifically identified. Control of information produced by the software once it has been transferred for operation is the responsibility of the receiving organization

  18. Gas characterization system software acceptance test procedure

    International Nuclear Information System (INIS)

    Vo, C.V.

    1996-01-01

    This document details the Software Acceptance Testing of gas characterization systems. The gas characterization systems will be used to monitor the vapor spaces of waste tanks known to contain measurable concentrations of flammable gases

  19. Software Intensive Systems Cost and Schedule Estimation

    Science.gov (United States)

    2013-06-13

    of labor counted in or across each activity. The activity data in  the SRDR is reported following the [ ISO   12207 ] processes for software development...Release  Table 19 ISO /IEC 12207 Development Activities System requirements analysis System architectural design A ct iv iti es in S RD R da ta... 12207 ]  ISO /IEC  12207 , International Standard on Information Technology  Software Lifecycle Processes, International Organization for  Standardization

  20. Software hazard analysis for nuclear digital protection system by Colored Petri Net

    International Nuclear Information System (INIS)

    Bai, Tao; Chen, Wei-Hua; Liu, Zhen; Gao, Feng

    2017-01-01

    Highlights: •A dynamic hazard analysis method is proposed for the safety-critical software. •The mechanism relies on Colored Petri Net. •Complex interactions between software and hardware are captured properly. •Common failure mode in software are identified effectively. -- Abstract: The software safety of a nuclear digital protection system is critical for the safety of nuclear power plants as any software defect may result in severe damage. In order to ensure the safety and reliability of safety-critical digital system products and their applications, software hazard analysis is required to be performed during the lifecycle of software development. The dynamic software hazard modeling and analysis method based on Colored Petri Net is proposed and applied to the safety-critical control software of the nuclear digital protection system in this paper. The analysis results show that the proposed method can explain the complex interactions between software and hardware and identify the potential common cause failure in software properly and effectively. Moreover, the method can find the dominant software induced hazard to safety control actions, which aids in increasing software quality.

  1. A Novel Technique for Running the NASA Legacy Code LAPIN Synchronously With Simulations Developed Using Simulink

    Science.gov (United States)

    Vrnak, Daniel R.; Stueber, Thomas J.; Le, Dzu K.

    2012-01-01

    This report presents a method for running a dynamic legacy inlet simulation in concert with another dynamic simulation that uses a graphical interface. The legacy code, NASA's LArge Perturbation INlet (LAPIN) model, was coded using the FORTRAN 77 (The Portland Group, Lake Oswego, OR) programming language to run in a command shell similar to other applications that used the Microsoft Disk Operating System (MS-DOS) (Microsoft Corporation, Redmond, WA). Simulink (MathWorks, Natick, MA) is a dynamic simulation that runs on a modern graphical operating system. The product of this work has both simulations, LAPIN and Simulink, running synchronously on the same computer with periodic data exchanges. Implementing the method described in this paper avoided extensive changes to the legacy code and preserved its basic operating procedure. This paper presents a novel method that promotes inter-task data communication between the synchronously running processes.

  2. Hardware-assisted software clock synchronization for homogeneous distributed systems

    Science.gov (United States)

    Ramanathan, P.; Kandlur, Dilip D.; Shin, Kang G.

    1990-01-01

    A clock synchronization scheme that strikes a balance between hardware and software solutions is proposed. The proposed is a software algorithm that uses minimal additional hardware to achieve reasonably tight synchronization. Unlike other software solutions, the guaranteed worst-case skews can be made insensitive to the maximum variation of message transit delay in the system. The scheme is particularly suitable for large partially connected distributed systems with topologies that support simple point-to-point broadcast algorithms. Examples of such topologies include the hypercube and the mesh interconnection structures.

  3. (Re)surveying Mediterranean rural landscapes : GIS and legacy survey data.

    OpenAIRE

    Witcher, R. E.

    2008-01-01

    Legacy data have always been important for Mediterranean archaeologists. Over the past decade, one specific category of legacy data, that deriving from regional survey, has become particularly important. Not only has the scale of research questions become larger (requiring greater reliance on others' data), but the surface archaeological record is deteriorating (diminishing the ability to recover good data). The legacy data from many individual surveys have now been subject to digitisation an...

  4. Integrated analysis software for bulk power system stability

    Energy Technology Data Exchange (ETDEWEB)

    Tanaka, T; Nagao, T; Takahashi, K [Central Research Inst. of Electric Power Industry, Tokyo (Japan)

    1994-12-31

    This paper presents Central Research Inst.of Electric Power Industry - CRIEPI`s - own developed three softwares for bulk power network analysis and the user support system which arranges tremendous data necessary for these softwares with easy and high reliability. (author) 3 refs., 7 figs., 2 tabs.

  5. The Legacy Project--William E. Dugger, Jr., DTE

    Science.gov (United States)

    Moye, Johnny J.; Dugger, William E., Jr.

    2016-01-01

    This is the ninth in a series of articles entitled "The Legacy Project." The Legacy Project focuses on the lives and actions of leaders who have forged the educator profession into what it is today. Members of the profession owe a debt of gratitude to these leaders. One simple way to demonstrate that gratitude is to recognize these…

  6. Gas characterization system software acceptance test report

    International Nuclear Information System (INIS)

    Vo, C.V.

    1996-01-01

    This document details the results of software acceptance testing of gas characterization systems. The gas characterization systems will be used to monitor the vapor spaces of waste tanks known to contain measurable concentrations of flammable gases

  7. Catchment legacies and time lags: a parsimonious watershed model to predict the effects of legacy storage on nitrogen export.

    Directory of Open Access Journals (Sweden)

    Kimberly J Van Meter

    Full Text Available Nutrient legacies in anthropogenic landscapes, accumulated over decades of fertilizer application, lead to time lags between implementation of conservation measures and improvements in water quality. Quantification of such time lags has remained difficult, however, due to an incomplete understanding of controls on nutrient depletion trajectories after changes in land-use or management practices. In this study, we have developed a parsimonious watershed model for quantifying catchment-scale time lags based on both soil nutrient accumulations (biogeochemical legacy and groundwater travel time distributions (hydrologic legacy. The model accurately predicted the time lags observed in an Iowa watershed that had undergone a 41% conversion of area from row crop to native prairie. We explored the time scales of change for stream nutrient concentrations as a function of both natural and anthropogenic controls, from topography to spatial patterns of land-use change. Our results demonstrate that the existence of biogeochemical nutrient legacies increases time lags beyond those due to hydrologic legacy alone. In addition, we show that the maximum concentration reduction benefits vary according to the spatial pattern of intervention, with preferential conversion of land parcels having the shortest catchment-scale travel times providing proportionally greater concentration reductions as well as faster response times. In contrast, a random pattern of conversion results in a 1:1 relationship between percent land conversion and percent concentration reduction, irrespective of denitrification rates within the landscape. Our modeling framework allows for the quantification of tradeoffs between costs associated with implementation of conservation measures and the time needed to see the desired concentration reductions, making it of great value to decision makers regarding optimal implementation of watershed conservation measures.

  8. Creating legacy through evaluation

    DEFF Research Database (Denmark)

    Degn, Hans-Peter; Lynghøj, Hanne; Hansen, Louise Ejgod

    Contemporary discussions regarding sustainability and cultural policy increasingly tends to focus on the longterm perspective of cultural legacy. This paper addresses the complex relation between an overall program and its underlying projects and activities. A central question in this respect is ...

  9. Unified Engineering Software System

    Science.gov (United States)

    Purves, L. R.; Gordon, S.; Peltzman, A.; Dube, M.

    1989-01-01

    Collection of computer programs performs diverse functions in prototype engineering. NEXUS, NASA Engineering Extendible Unified Software system, is research set of computer programs designed to support full sequence of activities encountered in NASA engineering projects. Sequence spans preliminary design, design analysis, detailed design, manufacturing, assembly, and testing. Primarily addresses process of prototype engineering, task of getting single or small number of copies of product to work. Written in FORTRAN 77 and PROLOG.

  10. Legacy Sediments in U.S. River Environments: Atrazine and Aggradation to Zinc and Zoobenthos

    Science.gov (United States)

    Wohl, E.

    2014-12-01

    Legacy sediments are those that are altered by human activities. Alterations include (i) human-caused aggradation (and subsequent erosion), such as sediment accumulating upstream from relict or contemporary dams, (ii) human-caused lack of continuing deposition that results in changing moisture and nutrient levels within existing sediments, such as on floodplains that no longer receive lateral or vertical accretion deposits because of levees, bank stabilization, and other channel engineering, and (iii) human-generated contaminants such as PCBs and pesticides that adsorb to fine sediment. Existing estimates of human alterations of river systems suggest that legacy sediments are ubiquitous. Only an estimated 2% of river miles in the United States are not affected by flow regulation that alters sediment transport, for example, and less than half of major river basins around the world are minimally altered by flow regulation. Combined with extensive but poorly documented reduction in floodplain sedimentation, as well as sediment contamination by diverse synthetic compounds, excess nutrients, and heavy metals, these national and global estimates suggest that legacy sediments now likely constitute a very abundant type of fluvial sediment. Because legacy sediments can alter river form and function for decades to centuries after the cessation of the human activity that created the legacy sediments, river management and restoration must be informed by accurate knowledge of the distribution and characteristics of legacy sediments. Geomorphologists can contribute understanding of sediment dynamics, including: the magnitude, frequency, and duration of flows that mobilize sediments with adsorbed contaminants; sites where erosion and deposition are most likely to occur under specified flow and sediment supply; residence time of sediments; and the influence of surface and subsurface water fluxes on sediment stability and geochemistry.

  11. Software Systems for High-performance Quantum Computing

    Energy Technology Data Exchange (ETDEWEB)

    Humble, Travis S [ORNL; Britt, Keith A [ORNL

    2016-01-01

    Quantum computing promises new opportunities for solving hard computational problems, but harnessing this novelty requires breakthrough concepts in the design, operation, and application of computing systems. We define some of the challenges facing the development of quantum computing systems as well as software-based approaches that can be used to overcome these challenges. Following a brief overview of the state of the art, we present models for the quantum programming and execution models, the development of architectures for hybrid high-performance computing systems, and the realization of software stacks for quantum networking. This leads to a discussion of the role that conventional computing plays in the quantum paradigm and how some of the current challenges for exascale computing overlap with those facing quantum computing.

  12. RAVONSICS-challenging for assuring software reliability of nuclear I and C system

    International Nuclear Information System (INIS)

    Hai Zeng; Ming Yang; Yoshikawa, Hidekazu

    2015-01-01

    As the “central nerve system”, the highly reliable Instrumentation and Control (I and C) systems, which provide the right functions and functions correctly, are always desirable not only for the end users of NPPs but also the suppliers of I and C systems. The Digitalization of nuclear I and C system happened in recent years brought a lot of new features for nuclear I and C system. On one side digital technology provides more functionalities, and it should be more reliable and robust; on the other side, digital technology brings new challenge for nuclear I and C system, especially the software running in the hardware component. The software provides flexible functionalities for nuclear I and C system, but it also brings the difficulties to evaluate the reliability and safety of it because of the complexity of software. The reliability of software, which is indispensable part of I and C system, will have essential impact on the reliability of the whole system, and people definitely want to know what the reliability of this intangible part is. The methods used for the evaluation of reliability of system and hardware hardly work for software, because the inherent difference of failure mechanism exists between software and hardware. Failure in software is systematically induced by design error, but failure in hardware is randomly induced by material and production. To continue the effort on this hot topic and to try to achieve consensus on the potential methodology for software reliability evaluation, a cooperative research project called RAVONSICS (Reliability and Verification and Validation of Nuclear Safety I and C Software) is being carried on by 7 Chinese partners, which includes University, research institute, utility, vendor, and safety regulatory body. The objective of RAVONSICS is to bring forwards the methodology for the software reliability evaluation, and the software verification technique. RAVONSICS works cooperatively with its European sister project

  13. Hotel software-comprehensive hotel systems

    OpenAIRE

    Šilhová, Lenka

    2010-01-01

    This bachelor's thesis deals with the usage of computer systems in the hotel industry. First part is focused on history, development and integration of technology into this field. Second part is dedicated to concrete products of the company Micros-Fidelio, which is the leader of hotel software market in the Czech Republic.

  14. Coordination Approaches for Complex Software Systems

    NARCIS (Netherlands)

    Bosse, T.; Hoogendoorn, M.; Treur, J.

    2006-01-01

    This document presents the results of a collaboration between the Vrije Universiteit Amsterdam, Department of Artificial Intelligence and Force Vision to investigate coordination approaches for complex software systems. The project was funded by Force Vision.

  15. Archiving Software Systems: Approaches to Preserve Computational Capabilities

    Science.gov (United States)

    King, T. A.

    2014-12-01

    A great deal of effort is made to preserve scientific data. Not only because data is knowledge, but it is often costly to acquire and is sometimes collected under unique circumstances. Another part of the science enterprise is the development of software to process and analyze the data. Developed software is also a large investment and worthy of preservation. However, the long term preservation of software presents some challenges. Software often requires a specific technology stack to operate. This can include software, operating systems and hardware dependencies. One past approach to preserve computational capabilities is to maintain ancient hardware long past its typical viability. On an archive horizon of 100 years, this is not feasible. Another approach to preserve computational capabilities is to archive source code. While this can preserve details of the implementation and algorithms, it may not be possible to reproduce the technology stack needed to compile and run the resulting applications. This future forward dilemma has a solution. Technology used to create clouds and process big data can also be used to archive and preserve computational capabilities. We explore how basic hardware, virtual machines, containers and appropriate metadata can be used to preserve computational capabilities and to archive functional software systems. In conjunction with data archives, this provides scientist with both the data and capability to reproduce the processing and analysis used to generate past scientific results.

  16. Software reliability growth model for safety systems of nuclear reactor

    International Nuclear Information System (INIS)

    Thirugnana Murthy, D.; Murali, N.; Sridevi, T.; Satya Murty, S.A.V.; Velusamy, K.

    2014-01-01

    The demand for complex software systems has increased more rapidly than the ability to design, implement, test, and maintain them, and the reliability of software systems has become a major concern for our, modern society.Software failures have impaired several high visibility programs in space, telecommunications, defense and health industries. Besides the costs involved, it setback the projects. The ways of quantifying it and using it for improvement and control of the software development and maintenance process. This paper discusses need for systematic approaches for measuring and assuring software reliability which is a major share of project development resources. It covers the reliability models with the concern on 'Reliability Growth'. It includes data collection on reliability, statistical estimation and prediction, metrics and attributes of product architecture, design, software development, and the operational environment. Besides its use for operational decisions like deployment, it includes guiding software architecture, development, testing and verification and validation. (author)

  17. Methodologic model to scheduling on service systems: a software engineering approach

    Directory of Open Access Journals (Sweden)

    Eduyn Ramiro Lopez-Santana

    2016-06-01

    Full Text Available This paper presents an approach of software engineering to a research proposal to make an Expert System to scheduling on service systems using methodologies and processes of software development. We use the adaptive software development as methodology for the software architecture based on the description as a software metaprocess that characterizes the research process. We make UML’s diagrams (Unified Modeling Language to provide a visual modeling that describes the research methodology in order to identify the actors, elements and interactions in the research process.

  18. Honeywell modular automation system computer software documentation

    International Nuclear Information System (INIS)

    Cunningham, L.T.

    1997-01-01

    This document provides a Computer Software Documentation for a new Honeywell Modular Automation System (MAS) being installed in the Plutonium Finishing Plant (PFP). This system will be used to control new thermal stabilization furnaces in HA-21I

  19. Use of Commericially Available Software in an Attribute Measurement System

    International Nuclear Information System (INIS)

    MacArthur, Duncan W.; Bracken, David S.; Carrillo, Louis A.; Elmont, Timothy H.; Frame, Katherine C.; Hirsch, Karen L.

    2005-01-01

    A major issue in international safeguards of nuclear materials is the ability to verify that processes and materials in nuclear facilities are consistent with declaration without revealing sensitive information. An attribute measurement system (AMS) is a non-destructive assay (NDA) system that utilizes an information barrier to protect potentially sensitive information about the measurement item. A key component is the software utilized for operator interface, data collection, analysis, and attribute determination, as well as the operating system under which they are implemented. Historically, custom software has been used almost exclusively in transparency applications, and it is unavoidable that some amount of custom software is needed. The focus of this paper is to explore the extent to which commercially available software may be used and the relative merits.

  20. Programming Guidelines for FBD Programs in Reactor Protection System Software

    International Nuclear Information System (INIS)

    Jung, Se Jin; Lee, Dong Ah; Kim, Eui Sub; Yoo, Jun Beom; Lee, Jang Su

    2014-01-01

    Properties of programming languages, such as reliability, traceability, etc., play important roles in software development to improve safety. Several researches are proposed guidelines about programming to increase the dependability of software which is developed for safety critical systems. Misra-c is a widely accepted programming guidelines for the C language especially in the sector of vehicle industry. NUREG/CR-6463 helps engineers in nuclear industry develop software in nuclear power plant systems more dependably. FBD (Function Block Diagram), which is one of programming languages defined in IEC 61131-3 standard, is often used for software development of PLC (programmable logic controllers) in nuclear power plants. Software development for critical systems using FBD needs strict guidelines, because FBD is a general language and has easily mistakable elements. There are researches about guidelines for IEC 61131-3 programming languages. They, however, do not specify details about how to use languages. This paper proposes new guidelines for the FBD based on NUREG/CR-6463. The paper introduces a CASE (Computer-Aided Software Engineering) tool to check FBD programs with the new guidelines and shows availability with a case study using a FBD program in a reactor protection system. The paper is organized as follows

  1. Programming Guidelines for FBD Programs in Reactor Protection System Software

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Se Jin; Lee, Dong Ah; Kim, Eui Sub; Yoo, Jun Beom [Division of Computer Science and Engineering College of Information and Communication, Konkuk University, Seoul (Korea, Republic of); Lee, Jang Su [Man-Machine Interface System team Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2014-10-15

    Properties of programming languages, such as reliability, traceability, etc., play important roles in software development to improve safety. Several researches are proposed guidelines about programming to increase the dependability of software which is developed for safety critical systems. Misra-c is a widely accepted programming guidelines for the C language especially in the sector of vehicle industry. NUREG/CR-6463 helps engineers in nuclear industry develop software in nuclear power plant systems more dependably. FBD (Function Block Diagram), which is one of programming languages defined in IEC 61131-3 standard, is often used for software development of PLC (programmable logic controllers) in nuclear power plants. Software development for critical systems using FBD needs strict guidelines, because FBD is a general language and has easily mistakable elements. There are researches about guidelines for IEC 61131-3 programming languages. They, however, do not specify details about how to use languages. This paper proposes new guidelines for the FBD based on NUREG/CR-6463. The paper introduces a CASE (Computer-Aided Software Engineering) tool to check FBD programs with the new guidelines and shows availability with a case study using a FBD program in a reactor protection system. The paper is organized as follows.

  2. SAGA: A project to automate the management of software production systems

    Science.gov (United States)

    Campbell, Roy H.; Laliberte, D.; Render, H.; Sum, R.; Smith, W.; Terwilliger, R.

    1987-01-01

    The Software Automation, Generation and Administration (SAGA) project is investigating the design and construction of practical software engineering environments for developing and maintaining aerospace systems and applications software. The research includes the practical organization of the software lifecycle, configuration management, software requirements specifications, executable specifications, design methodologies, programming, verification, validation and testing, version control, maintenance, the reuse of software, software libraries, documentation, and automated management.

  3. Software System for the Calibration of X-Ray Measuring Instruments

    International Nuclear Information System (INIS)

    Gaytan-Gallardo, E.; Tovar-Munoz, V. M.; Cruz-Estrada, P.; Vergara-Martinez, F. J.; Rivero-Gutierrez, T.

    2006-01-01

    A software system that facilities the calibration of X-ray measuring instruments used in medical applications is presented. The Secondary Standard Dosimetry Laboratory (SSDL) of the Nuclear Research National Institute in Mexico (ININ in Spanish), supports activities concerning with ionizing radiations in medical area. One of these activities is the calibration of X-ray measuring instruments, in terms of air kerma or exposure by substitution method in an X-ray beam at a point where the rate has been determined by means of a standard ionization chamber. To automatize this process, a software system has been developed, the calibration system is composed by an X-ray unit, a Dynalizer IIIU X-ray meter by RADCAL, a commercial data acquisition card, the software system and the units to be tested and calibrated. A quality control plan has been applied in the development of the software system, ensuring that quality assurance procedures and standards are being followed

  4. Cesium legacy safety project management work plan

    International Nuclear Information System (INIS)

    Durham, J.S.

    1998-01-01

    This Management Work Plan (MWP) describes the process flow, quality assurance controls, and the Environment, Safety, and Health requirements of the Cesium Legacy Safety Project. This MWP provides an overview of the project goals and methods for repackaging the non-conforming Type W overpacks and packaging the CsCl powder and pellets. This MWP is not intended to apply to other activities associated with the CsCl Legacy Safety Program (i.e., clean out of South Cell)

  5. Practicality for Software Hazard Analysis for Nuclear Safety I and C System

    International Nuclear Information System (INIS)

    Kim, Yong-Ho; Moon, Kwon-Ki; Chang, Young-Woo; Jeong, Soo-Hyun

    2016-01-01

    We are using the concept of system safety in engineering. It is difficult to make any system perfectly safe and probably a complete system may not easily be achieved. The standard definition of a system from MIL-STD- 882E is: “The organization of hardware, software, material, facilities, personnel, data, and services needed to perform a designated function within a stated environment with specified results.” From the perspective of the system safety engineer and the hazard analysis process, software is considered as a subsystem. Regarding hazard analysis, to date, methods for identifying software failures and determining their effects is still a research problem. Since the success of software development is based on rigorous test of hardware and software, it is necessary to check the balance between software test and hardware test, and in terms of efficiency. Lessons learned and experience from similar systems are important for the work of hazard analysis. No major hazard has been issued for the software developed and verified in Korean NPPs. In addition to hazard analysis, software development, and verification and validation were thoroughly performed. It is reasonable that the test implementation including the development of the test case, stress and abnormal conditions, error recovery situations, and high risk hazardous situations play a key role in detecting and preventing software faults

  6. Practicality for Software Hazard Analysis for Nuclear Safety I and C System

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Yong-Ho; Moon, Kwon-Ki; Chang, Young-Woo; Jeong, Soo-Hyun [KEPCO Engineering and Construction Co., Deajeon (Korea, Republic of)

    2016-10-15

    We are using the concept of system safety in engineering. It is difficult to make any system perfectly safe and probably a complete system may not easily be achieved. The standard definition of a system from MIL-STD- 882E is: “The organization of hardware, software, material, facilities, personnel, data, and services needed to perform a designated function within a stated environment with specified results.” From the perspective of the system safety engineer and the hazard analysis process, software is considered as a subsystem. Regarding hazard analysis, to date, methods for identifying software failures and determining their effects is still a research problem. Since the success of software development is based on rigorous test of hardware and software, it is necessary to check the balance between software test and hardware test, and in terms of efficiency. Lessons learned and experience from similar systems are important for the work of hazard analysis. No major hazard has been issued for the software developed and verified in Korean NPPs. In addition to hazard analysis, software development, and verification and validation were thoroughly performed. It is reasonable that the test implementation including the development of the test case, stress and abnormal conditions, error recovery situations, and high risk hazardous situations play a key role in detecting and preventing software faults.

  7. Sustainable legacies for the 2012 Olympic Games.

    Science.gov (United States)

    Shipway, Richard

    2007-05-01

    The London 2012 Olympic and Paralympic Games have the unique potential to deliver sustainable sporting, social, cultural, economic and environmental legacies, not just for London as the host city, but for the whole of Britain. This article focuses primarily on the first three of these potential Olympics legacies. The first area explored is the social legacy as it impacts on host communities; second, the potential educational and cultural legacy of the 2012 Games are examined; and finally, there follows an overview of the health benefits that could result from a sustained increase in mass participation in sport, physical activity and exercise. This appraisal is undertaken through a review of existing Olympic literature and examples are drawn from previous summer and winter Games. This preliminary exploration is followed by the identification of some key challenges to be overcome if the opportunities available to a wide and diverse range of stakeholders are to be fully optimized. The article suggests that the 2012 Games can act as a catalyst for sports development throughout Britain, while also assisting with government cross-cutting agendas such as tackling crime, antisocial behaviour, developing healthy and active communities, improving educational attainment, and combating barriers to participation. In doing so, this article argues that priority should be placed at supporting grassroots sport through greater access to sport in the community, and not solely elite level sports development. The article concludes by suggesting that the 2012 Games provide opportunities to deliver real and tangible changes and most importantly, to afford a higher priority to sport, along with the obvious associated health benefits for Britain as a whole. The underlying challenge as we move towards 2012 is to achieve a positive step change in the attitudes towards sport and physical activity in British society. Achieving this would possibly be the greatest legacy of the 2012 Olympic and

  8. Progressive retry for software error recovery in distributed systems

    Science.gov (United States)

    Wang, Yi-Min; Huang, Yennun; Fuchs, W. K.

    1993-01-01

    In this paper, we describe a method of execution retry for bypassing software errors based on checkpointing, rollback, message reordering and replaying. We demonstrate how rollback techniques, previously developed for transient hardware failure recovery, can also be used to recover from software faults by exploiting message reordering to bypass software errors. Our approach intentionally increases the degree of nondeterminism and the scope of rollback when a previous retry fails. Examples from our experience with telecommunications software systems illustrate the benefits of the scheme.

  9. Legacy effects of drought on plant growth and the soil food web

    DEFF Research Database (Denmark)

    de Vries, Franciska; Liiri, Mira; Strandmark, Lisa Bjørnlund

    2012-01-01

    the potential to feed back on each other's performance. In a greenhouse experiment, we compared legacy effects of repeated drought on plant growth and the soil food web in two contrasting land-use systems: extensively managed grassland, rich in C and with a fungal-based food web, and intensively managed wheat...... lower in C and with a bacterial-based food web. Moreover, we assessed the effect of plant presence on the recovery of the soil food web after drought. Drought legacy effects increased plant growth in both systems, and a plant strongly reduced N leaching. Fungi, bacteria, and their predators were more...... resilient after drought in the grassland soil than in the wheat soil. The presence of a plant strongly affected the composition of the soil food web, and alleviated the effects of drought for most trophic groups, regardless of the system. This effect was stronger for the bottom trophic levels, whose...

  10. Quantitative reliability assessment for safety critical system software

    International Nuclear Information System (INIS)

    Chung, Dae Won; Kwon, Soon Man

    2005-01-01

    An essential issue in the replacement of the old analogue I and C to computer-based digital systems in nuclear power plants is the quantitative software reliability assessment. Software reliability models have been successfully applied to many industrial applications, but have the unfortunate drawback of requiring data from which one can formulate a model. Software which is developed for safety critical applications is frequently unable to produce such data for at least two reasons. First, the software is frequently one-of-a-kind, and second, it rarely fails. Safety critical software is normally expected to pass every unit test producing precious little failure data. The basic premise of the rare events approach is that well-tested software does not fail under normal routine and input signals, which means that failures must be triggered by unusual input data and computer states. The failure data found under the reasonable testing cases and testing time for these conditions should be considered for the quantitative reliability assessment. We will present the quantitative reliability assessment methodology of safety critical software for rare failure cases in this paper

  11. The Effect of Superstar Software on Hardware Sales in System Markets

    OpenAIRE

    Binken, Jeroen; Stremersch, Stefan

    2008-01-01

    textabstractSystems are composed of complementary products (e.g., video game systems are composed of the video game console and video games). Prior literature on indirect network effects argues that, in system markets, sales of the primary product (often referred to as "hardware") largely depend on the availability of complementary products (often referred to as "software"). Mathematical and empirical analyses have almost exclusively operationalized software availability as software quantity....

  12. A wide spectrum type system for transformation theory

    OpenAIRE

    Ladkau, Matthias

    2009-01-01

    One of the most difficult tasks a programmer can be confronted with is the migration of a legacy system. Usually, these systems are unstructured, poorly documented and contain complex program logic. The reason for this, in most cases, is an emphasis on raw performance rather than on clean and structured code as well as a long period of applying quick fixes and enhancements rather than doing a proper software reengineering process including a full redesign during major enhancements. Nowadays, ...

  13. Fielding a structural health monitoring system on legacy military aircraft: A business perspective

    Energy Technology Data Exchange (ETDEWEB)

    Bos, Marcel J. [Dept. of Gas Turbines and Structural Integrity, National Aerospace Laboratory NLR, Amsterdam (Netherlands)

    2015-12-15

    An important trend in the sustainment of military aircraft is the transition from preventative maintenance to condition based maintenance (CBM). For CBM, it is essential that the actual system condition can be measured and the measured condition can be reliably extrapolated to a convenient moment in the future in order to facilitate the planning process while maintaining flight safety. Much research effort is currently being made for the development of technologies that enable CBM, including structural health monitoring (SHM) systems. Great progress has already been made in sensors, sensor networks, data acquisition, models and algorithms, data fusion/mining techniques, etc. However, the transition of these technologies into service is very slow. This is because business cases are difficult to define and the certification of the SHM systems is very challenging. This paper describes a possibility for fielding a SHM system on legacy military aircraft with a minimum amount of certification issues and with a good prospect of a positive return on investment. For appropriate areas in the airframe the application of SHM will reconcile the fail-safety and slow crack growth damage tolerance approaches that can be used for safeguarding the continuing airworthiness of these areas, combining the benefits of both approaches and eliminating the drawbacks.

  14. Fielding a structural health monitoring system on legacy military aircraft: A business perspective

    International Nuclear Information System (INIS)

    Bos, Marcel J.

    2015-01-01

    An important trend in the sustainment of military aircraft is the transition from preventative maintenance to condition based maintenance (CBM). For CBM, it is essential that the actual system condition can be measured and the measured condition can be reliably extrapolated to a convenient moment in the future in order to facilitate the planning process while maintaining flight safety. Much research effort is currently being made for the development of technologies that enable CBM, including structural health monitoring (SHM) systems. Great progress has already been made in sensors, sensor networks, data acquisition, models and algorithms, data fusion/mining techniques, etc. However, the transition of these technologies into service is very slow. This is because business cases are difficult to define and the certification of the SHM systems is very challenging. This paper describes a possibility for fielding a SHM system on legacy military aircraft with a minimum amount of certification issues and with a good prospect of a positive return on investment. For appropriate areas in the airframe the application of SHM will reconcile the fail-safety and slow crack growth damage tolerance approaches that can be used for safeguarding the continuing airworthiness of these areas, combining the benefits of both approaches and eliminating the drawbacks

  15. Evaluating software for safety systems in nuclear power plants

    International Nuclear Information System (INIS)

    Lawrence, J.D.; Persons, W.L.; Preckshot, G.G.; Gallagher, J.

    1994-01-01

    In 1991, LLNL was asked by the NRC to provide technical assistance in various aspects of computer technology that apply to computer-based reactor protection systems. This has involved the review of safety aspects of new reactor designs and the provision of technical advice on the use of computer technology in systems important to reactor safety. The latter includes determining and documenting state-of-the-art subjects that require regulatory involvement by the NRC because of their importance in the development and implementation of digital computer safety systems. These subjects include data communications, formal methods, testing, software hazards analysis, verification and validation, computer security, performance, software complexity and others. One topic software reliability and safety is the subject of this paper

  16. Software for ASS-500 based early warning system

    International Nuclear Information System (INIS)

    Lipinski, P.; Isajenko, K.

    1998-01-01

    The article describes the software for the management of early warning system based on ASS-500 station. The software can communicate with the central computer using TCP/IP protocol. This allows remote control of the station through modem or local area network connection. The article describes Windows based user interface of the program

  17. Features of the Upgraded Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) Software

    Science.gov (United States)

    Mason, Michelle L.; Rufer, Shann J.

    2016-01-01

    The Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) software is used at the NASA Langley Research Center to analyze global aeroheating data on wind tunnel models tested in the Langley Aerothermodynamics Laboratory. One-dimensional, semi-infinite heating data derived from IHEAT are used in the design of thermal protection systems for hypersonic vehicles that are exposed to severe aeroheating loads, such as reentry vehicles during descent and landing procedures. This software program originally was written in the PV-WAVE(Registered Trademark) programming language to analyze phosphor thermography data from the two-color, relative-intensity system developed at Langley. To increase the efficiency, functionality, and reliability of IHEAT, the program was migrated to MATLAB(Registered Trademark) syntax and compiled as a stand-alone executable file labeled version 4.0. New features of IHEAT 4.0 include the options to perform diagnostic checks of the accuracy of the acquired data during a wind tunnel test, to extract data along a specified multi-segment line following a feature such as a leading edge or a streamline, and to batch process all of the temporal frame data from a wind tunnel run. Results from IHEAT 4.0 were compared on a pixel level to the output images from the legacy software to validate the program. The absolute differences between the heat transfer data output from the two programs were on the order of 10(exp -5) to 10(exp -7). IHEAT 4.0 replaces the PV-WAVE(Registered Trademark) version as the production software for aeroheating experiments conducted in the hypersonic facilities at NASA Langley.

  18. Software Safety Life cycle and Method of POSAFE-Q System

    International Nuclear Information System (INIS)

    Lee, Jang-Soo; Kwon, Kee-Choon

    2006-01-01

    This paper describes the relationship between the overall safety life cycle and the software safety life cycle during the development of the software based safety systems of Nuclear Power Plants. This includes the design and evaluation activities of components as well as the system. The paper also compares the safety life cycle and planning activities defined in IEC 61508 with those in IEC 60880, IEEE 7-4.3.2, and IEEE 1228. Using the KNICS project as an example, software safety life cycle and safety analysis methods applied to the POSAFE-Q are demonstrated. KNICS software safety life cycle is described by comparing to the software development, testing, and safety analysis process with international standards. The safety assessment of the software for POSAFE-Q is a joint Korean German project. The assessment methods applied in the project and the experiences gained from this project are presented

  19. Integrating open-source software applications to build molecular dynamics systems.

    Science.gov (United States)

    Allen, Bruce M; Predecki, Paul K; Kumosa, Maciej

    2014-04-05

    Three open-source applications, NanoEngineer-1, packmol, and mis2lmp are integrated using an open-source file format to quickly create molecular dynamics (MD) cells for simulation. The three software applications collectively make up the open-source software (OSS) suite known as MD Studio (MDS). The software is validated through software engineering practices and is verified through simulation of the diglycidyl ether of bisphenol-a and isophorone diamine (DGEBA/IPD) system. Multiple simulations are run using the MDS software to create MD cells, and the data generated are used to calculate density, bulk modulus, and glass transition temperature of the DGEBA/IPD system. Simulation results compare well with published experimental and numerical results. The MDS software prototype confirms that OSS applications can be analyzed against real-world research requirements and integrated to create a new capability. Copyright © 2014 Wiley Periodicals, Inc.

  20. Ground test accelerator control system software

    International Nuclear Information System (INIS)

    Burczyk, L.; Dalesio, R.; Dingler, R.; Hill, J.; Howell, J.A.; Kerstiens, D.; King, R.; Kozubal, A.; Little, C.; Martz, V.; Rothrock, R.; Sutton, J.

    1988-01-01

    This paper reports on the GTA control system that provides an environment in which the automation of a state-of-the-art accelerator can be developed. It makes use of commercially available computers, workstations, computer networks, industrial 110 equipment, and software. This system has built-in supervisory control (like most accelerator control systems), tools to support continuous control (like the process control industry), and sequential control for automatic start-up and fault recovery (like few other accelerator control systems). Several software tools support these levels of control: a real-time operating system (VxWorks) with a real-time kernel (VRTX), a configuration database, a sequencer, and a graphics editor. VxWorks supports multitasking, fast context-switching, and preemptive scheduling. VxWorks/VRTX is a network-based development environment specifically designed to work in partnership with the UNIX operating system. A data base provides the interface to the accelerator components. It consists of a run time library and a database configuration and editing tool. A sequencer initiates and controls the operation of all sequence programs (expressed as state programs). A graphics editor gives the user the ability to create color graphic displays showing the state of the machine in either text or graphics form

  1. Wings In Orbit: Scientific and Engineering Legacies of the Space Shuttle

    Science.gov (United States)

    Hale, N. Wayne (Editor); Lulla, Kamlesh (Editor); Lane, Helen W. (Editor); Chapline, Gail (Editor)

    2010-01-01

    This Space Shuttle book project reviews Wings In Orbit-scientific and engineering legacies of the Space Shuttle. The contents include: 1) Magnificent Flying Machine-A Cathedral to Technology; 2) The Historical Legacy; 3) The Shuttle and its Operations; 4) Engineering Innovations; 5) Major Scientific Discoveries; 6) Social, Cultural, and Educational Legacies; 7) Commercial Aerospace Industries and Spin-offs; and 8) The Shuttle continuum, Role of Human Spaceflight.

  2. The Utility of Open Source Software in Military Systems

    National Research Council Canada - National Science Library

    Esperon, Agustin I; Munoz, Jose P; Tanneau, Jean M

    2005-01-01

    .... The companies involved were THALES and GMV. The MILOS project aimed to demonstrate benefits of Open Source Software in large software based military systems, by casting off constraints inherent to traditional proprietary COTS and by taking...

  3. Modular Software for Spacecraft Navigation Using the Global Positioning System (GPS)

    Science.gov (United States)

    Truong, S. H.; Hartman, K. R.; Weidow, D. A.; Berry, D. L.; Oza, D. H.; Long, A. C.; Joyce, E.; Steger, W. L.

    1996-01-01

    The Goddard Space Flight Center Flight Dynamics and Mission Operations Divisions have jointly investigated the feasibility of engineering modular Global Positioning SYSTEM (GPS) navigation software to support both real time flight and ground postprocessing configurations. The goals of this effort are to define standard GPS data interfaces and to engineer standard, reusable navigation software components that can be used to build a broad range of GPS navigation support applications. The paper discusses the GPS modular software (GMOD) system and operations concepts, major requirements, candidate software architecture, feasibility assessment and recommended software interface standards. In additon, ongoing efforts to broaden the scope of the initial study and to develop modular software to support autonomous navigation using GPS are addressed,

  4. KAERI software safety guideline for developing safety-critical software in digital instrumentation and control system of nuclear power plant

    International Nuclear Information System (INIS)

    Lee, Jang Soo; Kim, Jang Yeol; Eum, Heung Seop.

    1997-07-01

    Recently, the safety planning for safety-critical software systems is being recognized as the most important phase in the software life cycle, and being developed new regulatory positions and standards by the regulatory and the standardization organization. The requirements for software important to safety of nuclear reactor are described in such positions and standards. Most of them are describing mandatory requirements, what shall be done, for the safety-critical software. The developers of such a software. However, there have been a lot of controversial factors on whether the work practices satisfy the regulatory requirements, and to justify the safety of such a system developed by the work practices, between the licenser and the licensee. We believe it is caused by the reason that there is a gap between the mandatory requirements (What) and the work practices (How). We have developed a guidance to fill such gap, which can be useful for both licenser and licensee to conduct a justification of the safety in the planning phase of developing the software for nuclear reactor protection systems. (author). 67 refs., 13 tabs., 2 figs

  5. KAERI software safety guideline for developing safety-critical software in digital instrumentation and control system of nuclear power plant

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Jang Soo; Kim, Jang Yeol; Eum, Heung Seop

    1997-07-01

    Recently, the safety planning for safety-critical software systems is being recognized as the most important phase in the software life cycle, and being developed new regulatory positions and standards by the regulatory and the standardization organization. The requirements for software important to safety of nuclear reactor are described in such positions and standards. Most of them are describing mandatory requirements, what shall be done, for the safety-critical software. The developers of such a software. However, there have been a lot of controversial factors on whether the work practices satisfy the regulatory requirements, and to justify the safety of such a system developed by the work practices, between the licenser and the licensee. We believe it is caused by the reason that there is a gap between the mandatory requirements (What) and the work practices (How). We have developed a guidance to fill such gap, which can be useful for both licenser and licensee to conduct a justification of the safety in the planning phase of developing the software for nuclear reactor protection systems. (author). 67 refs., 13 tabs., 2 figs.

  6. Conceptual Software Reliability Prediction Models for Nuclear Power Plant Safety Systems

    International Nuclear Information System (INIS)

    Johnson, G.; Lawrence, D.; Yu, H.

    2000-01-01

    The objective of this project is to develop a method to predict the potential reliability of software to be used in a digital system instrumentation and control system. The reliability prediction is to make use of existing measures of software reliability such as those described in IEEE Std 982 and 982.2. This prediction must be of sufficient accuracy to provide a value for uncertainty that could be used in a nuclear power plant probabilistic risk assessment (PRA). For the purposes of the project, reliability was defined to be the probability that the digital system will successfully perform its intended safety function (for the distribution of conditions under which it is expected to respond) upon demand with no unintended functions that might affect system safety. The ultimate objective is to use the identified measures to develop a method for predicting the potential quantitative reliability of a digital system. The reliability prediction models proposed in this report are conceptual in nature. That is, possible prediction techniques are proposed and trial models are built, but in order to become a useful tool for predicting reliability, the models must be tested, modified according to the results, and validated. Using methods outlined by this project, models could be constructed to develop reliability estimates for elements of software systems. This would require careful review and refinement of the models, development of model parameters from actual experience data or expert elicitation, and careful validation. By combining these reliability estimates (generated from the validated models for the constituent parts) in structural software models, the reliability of the software system could then be predicted. Modeling digital system reliability will also require that methods be developed for combining reliability estimates for hardware and software. System structural models must also be developed in order to predict system reliability based upon the reliability

  7. Tailorable software architectures in the accelerator control system environment

    International Nuclear Information System (INIS)

    Mejuev, Igor; Kumagai, Akira; Kadokura, Eiichi

    2001-01-01

    Tailoring is further evolution of an application after deployment in order to adapt it to requirements that were not accounted for in the original design. End-user tailorability has been extensively researched in applied computer science from HCl and software engineering perspectives. Tailorability allows coping with flexibility requirements, decreasing maintenance and development costs of software products. In general, dynamic or diverse software requirements constitute the need for implementing end-user tailorability in computer systems. In accelerator physics research the factor of dynamic requirements is especially important, due to frequent software and hardware modifications resulting in correspondingly high upgrade and maintenance costs. In this work we introduce the results of feasibility study on implementing end-user tailorability in the software for accelerator control system, considering the design and implementation of distributed monitoring application for 12 GeV KEK Proton Synchrotron as an example. The software prototypes used in this work are based on a generic tailoring platform (VEDICI), which allows decoupling of tailoring interfaces and runtime components. While representing a reusable application-independent framework, VEDICI can be potentially applied for tailoring of arbitrary compositional Web-based applications

  8. CheMentor Software System by H. A. Peoples

    Science.gov (United States)

    Reid, Brian P.

    1997-09-01

    CheMentor Software System H. A. Peoples. Computerized Learning Enhancements: http://www.ecis.com/~clehap; email: clehap@ecis.com; 1996 - 1997. CheMentor is a series of software packages for introductory-level chemistry, which includes Practice Items (I), Stoichiometry (I), Calculating Chemical Formulae, and the CheMentor Toolkit. The first three packages provide practice problems for students and various types of help to solve them; the Toolkit includes "calculators" for determining chemical quantities as well as the Practice Items (I) set of problems. The set of software packages is designed so that each individual product acts as a module of a common CheMentor program. As the name CheMentor implies, the software is designed as a "mentor" for students learning introductory chemistry concepts and problems. The typical use of the software would be by individual students (or perhaps small groups) as an adjunct to lectures. CheMentor is a HyperCard application and the modules are HyperCard stacks. The requirements to run the packages include a Macintosh computer with at least 1 MB of RAM, a hard drive with several MB of available space depending upon the packages selected (10 MB were required for all the packages reviewed here), and the Mac operating system 6.0.5 or later.

  9. SOFTM: a software maintenance expert system in Prolog

    DEFF Research Database (Denmark)

    Pau, L.; Negret, J. M.

    1988-01-01

    A description is given of a knowledge-based system called SOFTM, serving the following purposes: (1) assisting a software programmer or analyst in his application code maintenance tasks, (2) generating and updating automatically software correction documentation, (3) helping the end user register......, and on interfacing capabilities of Prolog II to a variety of other languages...

  10. Nuclear power plant monitoring and control system software: verification and validation

    International Nuclear Information System (INIS)

    Kaneda, M.; Niki, K.; Shibata, K.

    1986-01-01

    The design philosophy, configuration, and production of process computer system software used for the monitoring and control of nuclear power plants are presented in detail. To achieve a very complex software system that not only has excellent performance, high reliability, and full fail safe protection, but also is easy to produce, verify, and validate, and has flexibility for future modifications, we developed the following software production system to support safe operation of nuclear power stations. The fundamental design philosophy of our monitoring and control system software is the complete separation of program logic from the data base. The logic section is highly standardized and applicable to a wide range of power generation plant computer application systems. The plant-unique properties and characteristics are all described in the data base. This separation of logic and data base has a dramatic effect on the reliability and productivity of the software system. One of the main features of the data base system is the use of easy-to-learn, easy-to-use, problem-oriented language that enables non-programmers to build up the data base using simple fill-in-the-blank type tables. The generation of these tables is fully automated, and the full set of online table editing utility software, which runs on the target plant process computer, has proven very effective in incorporation of changes and modifications at the site. (author)

  11. Mining legacy data: digitizing well logs leads to bypassed pay opportunities

    Energy Technology Data Exchange (ETDEWEB)

    Polczer, S.

    1999-10-01

    International Datashare Corporation's (iDc) computer software that filters graphic data through an algorithm programmed to recognize pre-set patterns, a process similar to optical character recognition, is described. The software is particularly well suited to large volumes of legacy data because it does not recognize numbers or calculations. Instead, it scans well log files as high-resolution raster files, strips them of all numeric data to isolate hydrocarbon signatures. This pattern recognition is expected to expose (to trained professionals) significant bypassed oil and gas pay that at the time of first development of the well were considered worthless due to lack of markets and transportation infrastructure. iDc plans to distribute the raster well log library of nearly one million records by making it available online. iDc also intends to rasterize and index other data sets such as drill stem data and fluids data , and combine the information into a complete solutions package to mark the company's 50 years in business. This new 'data mining' application is intended to enable operators to locate and develop hydrocarbon reserves that were originally bypassed. 1 photo, 1 fig.

  12. Software architecture for the ORNL large-coil test facility data system

    International Nuclear Information System (INIS)

    Blair, E.T.; Baylor, L.R.

    1986-01-01

    The VAX-based data-acquisition system for the International Fusion Superconducting Magnet Test Facility (IFSMTF) at Oak Ridge National Laboratory (ORNL) is a second-generation system that evolved from a PDP-11/60-based system used during the initial phase of facility testing. The VAX-based software represents a layered implementation that provides integrated access to all of the data sources within the system, decoupling end-user data retrieval from various front-end data sources through a combination of software architecture and instrumentation data bases. Independent VAX processes manage the various front-end data sources, each being responsible for controlling, monitoring, acquiring, and disposing data and control parameters for access from the data retrieval software. This paper describes the software architecture and the functionality incorporated into the various layers of the data system

  13. Software architecture for the ORNL large coil test facility data system

    International Nuclear Information System (INIS)

    Blair, E.T.; Baylor, L.R.

    1986-01-01

    The VAX-based data acquisition system for the International Fusion Superconducting Magnet Test Facility (IFSMTF) at Oak Ridge National Laboratory (ORNL) is a second-generation system that evolved from a PDP-11/60-based system used during the initial phase of facility testing. The VAX-based software represents a layered implementation that provides integrated access to all of the data sources within the system, deoupling end-user data retrieval from various front-end data sources through a combination of software architecture and instrumentation data bases. Independent VAX processes manage the various front-end data sources, each being responsible for controlling, monitoring, acquiring and disposing data and control parameters for access from the data retrieval software. This paper describes the software architecture and the functionality incorporated into the various layers of the data system

  14. Software Reliability Issues Concerning Large and Safety Critical Software Systems

    Science.gov (United States)

    Kamel, Khaled; Brown, Barbara

    1996-01-01

    This research was undertaken to provide NASA with a survey of state-of-the-art techniques using in industrial and academia to provide safe, reliable, and maintainable software to drive large systems. Such systems must match the complexity and strict safety requirements of NASA's shuttle system. In particular, the Launch Processing System (LPS) is being considered for replacement. The LPS is responsible for monitoring and commanding the shuttle during test, repair, and launch phases. NASA built this system in the 1970's using mostly hardware techniques to provide for increased reliability, but it did so often using custom-built equipment, which has not been able to keep up with current technologies. This report surveys the major techniques used in industry and academia to ensure reliability in large and critical computer systems.

  15. Software control and system configuration management - A process that works

    Science.gov (United States)

    Petersen, K. L.; Flores, C., Jr.

    1983-01-01

    A comprehensive software control and system configuration management process for flight-crucial digital control systems of advanced aircraft has been developed and refined to insure efficient flight system development and safe flight operations. Because of the highly complex interactions among the hardware, software, and system elements of state-of-the-art digital flight control system designs, a systems-wide approach to configuration control and management has been used. Specific procedures are implemented to govern discrepancy reporting and reconciliation, software and hardware change control, systems verification and validation testing, and formal documentation requirements. An active and knowledgeable configuration control board reviews and approves all flight system configuration modifications and revalidation tests. This flexible process has proved effective during the development and flight testing of several research aircraft and remotely piloted research vehicles with digital flight control systems that ranged from relatively simple to highly complex, integrated mechanizations.

  16. Reliability estimation of safety-critical software-based systems using Bayesian networks

    International Nuclear Information System (INIS)

    Helminen, A.

    2001-06-01

    Due to the nature of software faults and the way they cause system failures new methods are needed for the safety and reliability evaluation of software-based safety-critical automation systems in nuclear power plants. In the research project 'Programmable automation system safety integrity assessment (PASSI)', belonging to the Finnish Nuclear Safety Research Programme (FINNUS, 1999-2002), various safety assessment methods and tools for software based systems are developed and evaluated. The project is financed together by the Radiation and Nuclear Safety Authority (STUK), the Ministry of Trade and Industry (KTM) and the Technical Research Centre of Finland (VTT). In this report the applicability of Bayesian networks to the reliability estimation of software-based systems is studied. The applicability is evaluated by building Bayesian network models for the systems of interest and performing simulations for these models. In the simulations hypothetical evidence is used for defining the parameter relations and for determining the ability to compensate disparate evidence in the models. Based on the experiences from modelling and simulations we are able to conclude that Bayesian networks provide a good method for the reliability estimation of software-based systems. (orig.)

  17. Advanced Transport Operating System (ATOPS) utility library software description

    Science.gov (United States)

    Clinedinst, Winston C.; Slominski, Christopher J.; Dickson, Richard W.; Wolverton, David A.

    1993-01-01

    The individual software processes used in the flight computers on-board the Advanced Transport Operating System (ATOPS) aircraft have many common functional elements. A library of commonly used software modules was created for general uses among the processes. The library includes modules for mathematical computations, data formatting, system database interfacing, and condition handling. The modules available in the library and their associated calling requirements are described.

  18. Biological field stations: research legacies and sites for serendipity

    Science.gov (United States)

    William K. Michener; Keith L. Bildstein; Arthur McKee; Robert R. Parmenter; William W. Hargrove; Deedra McClearn; Mark Stromberg

    2009-01-01

    Biological field stations are distributed throughout North America, capturing much of the ecological variability present at the continental scale and encompassing many unique habitats. In addition to their role in supporting research and education, field stations offer legacies of data, specimens, and accumulated knowledge. Such legacies often provide the only...

  19. WORLD CUP LEGACY AND PERTAINING IMPACTS ON SÃO PAULO CITY´S FUTURE

    Directory of Open Access Journals (Sweden)

    Marcos Vinicius Cardoso

    2013-08-01

    Full Text Available Legacies – structures that are built for events and which remain after the same - are one of the major positive aspects paraded by mega sporting events organizers. This study´s purpose is to analyze the current situation of legacies promised by the many governmental instances for the city of São Paulo - host city of Fifa´s 2014 World Cup – and prospect which legacies will become effective in the city. Preliminary assessments may raise construction concerns, alert the public to keep an eye on undertaken obligations and encourage official actions (Mangan, 2008, p. 1,871. Data was obtained from National Audit Court (TCU reports, Ministry and United Nations documents, in addition to testimonials and information gathered from some of Brazil´s major press media. Data analysis was conducted by classifying legacies according to tangible and intangible legacy concepts (Kaplanidou and Karadakis, 2010 followed by an analysis of promised legacies versus current status during the period of analysis. Finally, discussions as to most probable to come about legacies were presented. Results indicate that a portion of promised legacies stand a fair chance of achievement. On the other hand, other projects lag behind schedule or have been cancelled. Preliminary surveys suggest full completion of promised legacies is not possible, there has been an overuse of public resources as opposed to that planned, and provide indicatives as to the investment´s high opportunity cost.

  20. Cross-compilation of ATLAS online software to the power PC-Vx works system

    International Nuclear Information System (INIS)

    Tian Yuren; Li Jin; Ren Zhengyu; Zhu Kejun

    2005-01-01

    BES III, selected ATLAS online software as a framework of its run-control system. BES III applied Power PC-VxWorks system on its front-end readout system, so it is necessary to cross-compile this software to PowerPC-VxWorks system. The article demonstrates several aspects related to this project, such as the structure and organization of the ATLAS online software, the application of CMT tool while cross-compiling, the selection and configuration of the cross-compiler, methods to solve various problems due to the difference of compiler and operating system etc. The software, after cross-compiling, can normally run, and makes up a complete run-control system with the software running on Linux system. (authors)

  1. Additive Manufacturing: Which DLA-Managed Legacy Parts are Potential AM Candidates

    Science.gov (United States)

    2016-07-01

    R G ADDITIVE MANUFACTURING : WHICH DLA-MANAGED LEGACY PARTS ARE POTENTIAL AM CANDIDATES? REPORT DL501T1 J UL Y 2016...L Y 2 0 1 6 ADDITIVE MANUFACTURING : WHICH DLA-MANAGED LEGACY PARTS ARE POTENTIAL AM CANDIDATES? REPORT DL501T1 Thomas K . Pa rk s...DESIGNATED BY OTHER OFFICIAL DOCUMENTATION. LMI © 2016. ALL RIGHTS RESERVED. iii Additive Manufacturing : Which DLA-Managed Legacy Parts Are

  2. Software FMEA analysis for safety-related application software

    International Nuclear Information System (INIS)

    Park, Gee-Yong; Kim, Dong Hoon; Lee, Dong Young

    2014-01-01

    Highlights: • We develop a modified FMEA analysis suited for applying to software architecture. • A template for failure modes on a specific software language is established. • A detailed-level software FMEA analysis on nuclear safety software is presented. - Abstract: A method of a software safety analysis is described in this paper for safety-related application software. The target software system is a software code installed at an Automatic Test and Interface Processor (ATIP) in a digital reactor protection system (DRPS). For the ATIP software safety analysis, at first, an overall safety or hazard analysis is performed over the software architecture and modules, and then a detailed safety analysis based on the software FMEA (Failure Modes and Effect Analysis) method is applied to the ATIP program. For an efficient analysis, the software FMEA analysis is carried out based on the so-called failure-mode template extracted from the function blocks used in the function block diagram (FBD) for the ATIP software. The software safety analysis by the software FMEA analysis, being applied to the ATIP software code, which has been integrated and passed through a very rigorous system test procedure, is proven to be able to provide very valuable results (i.e., software defects) that could not be identified during various system tests

  3. A Hybrid Hardware and Software Component Architecture for Embedded System Design

    Science.gov (United States)

    Marcondes, Hugo; Fröhlich, Antônio Augusto

    Embedded systems are increasing in complexity, while several metrics such as time-to-market, reliability, safety and performance should be considered during the design of such systems. A component-based design which enables the migration of its components between hardware and software can cope to achieve such metrics. To enable that, we define hybrid hardware and software components as a development artifact that can be deployed by different combinations of hardware and software elements. In this paper, we present an architecture for developing such components in order to construct a repository of components that can migrate between the hardware and software domains to meet the design system requirements.

  4. Systems and software variability management concepts, tools and experiences

    CERN Document Server

    Capilla, Rafael; Kang, Kyo-Chul

    2013-01-01

    The success of product line engineering techniques in the last 15 years has popularized the use of software variability as a key modeling approach for describing the commonality and variability of systems at all stages of the software lifecycle. Software product lines enable a family of products to share a common core platform, while allowing for product specific functionality being built on top of the platform. Many companies have exploited the concept of software product lines to increase the resources that focus on highly differentiating functionality and thus improve their competitiveness

  5. A dependability modeling of software under hardware faults digitized system in nuclear power plants

    International Nuclear Information System (INIS)

    Choi, Jong Gyun

    1996-02-01

    An analytic approach to the dependability evaluation of software in the operational phase is suggested in this work with special attention to the physical fault effects on the software dependability : The physical faults considered are memory faults and the dependability measure in question is the reliability. The model is based on the simple reliability theory and the graph theory with the path decomposition micro model. The model represents an application software with a graph consisting of nodes and arcs that probabilistic ally determine the flow from node to node. Through proper transformation of nodes and arcs, the graph can be reduced to a simple two-node graph and the software failure probability is derived from this graph. This model can be extended to the software system which consists of several complete modules without modification. The derived model is validated by the computer simulation, where the software is transformed to a probabilistic control flow graph. Simulation also shows a different viewpoint of software failure behavior. Using this model, we predict the reliability of an application software and a software system in a digitized system(ILS system) in the nuclear power plant and show the sensitivity of the software reliability to the major physical parameters which affect the software failure in the normal operation phase. The derived model is validated by the computer simulation, where the software is transformed to a probabilistic control flow graph. Simulation also shows a different viewpoint of software failure behavior. Using this model, we predict the reliability of an application software and a software system in a digitized system (ILS system) is the nuclear power plant and show the sensitivity of the software reliability to the major physical parameters which affect the software failure in the normal operation phase. This modeling method is particularly attractive for medium size programs such as software used in digitized systems of

  6. Software Engineering Laboratory (SEL) data base reporting software user's guide and system description. Volume 1: Introduction and user's guide

    Science.gov (United States)

    1983-01-01

    Reporting software programs provide formatted listings and summary reports of the Software Engineering Laboratory (SEL) data base contents. The operating procedures and system information for 18 different reporting software programs are described. Sample output reports from each program are provided.

  7. Software system development of NPP plant DiD risk monitor. Basic design of software configuration

    International Nuclear Information System (INIS)

    Yoshikawa, Hidekazu; Nakagawa, Takashi

    2015-01-01

    A new risk monitor system is under development which can be applied not only to prevent severe accident in daily operation but also to serve as to mitigate the radiological hazard just after severe accident happens and long term management of post-severe accident consequences. The fundamental method for the new risk monitor system is first given on how to configure the Plant Defense in-Depth (DiD) Risk Monitor by object-oriented software system based on functional modeling approach. In this paper, software system for the plant DiD risk monitor is newly developed by object oriented method utilizing Unified Modeling Language (UML). Usage of the developed DiD risk monitor is also introduced by showing examples for LOCA case of AP1000. (author)

  8. VLSI systems energy management from a software perspective – A literature survey

    Directory of Open Access Journals (Sweden)

    Prasada Kumari K.S.

    2016-09-01

    Full Text Available The increasing demand for ultra-low power electronic systems has motivated research in device technology and hardware design techniques. Experimental studies have proved that the hardware innovations for power reduction are fully exploited only with the proper design of upper layer software. Also, the software power and energy modelling and analysis – the first step towards energy reduction is complex due to the inter and intra dependencies of processors, operating systems, application software, programming languages and compilers. The subject is too vast; the paper aims to give a consolidated view to researchers in arriving at solutions to power optimization problems from a software perspective. The review emphasizes the fact that software design and implementation is to be viewed from system energy conservation angle rather than as an isolated process. After covering a global view of end to end software based power reduction techniques for micro sensor nodes to High Performance Computing systems, specific design aspects related to battery powered Embedded computing for mobile and portable systems are addressed in detail. The findings are consolidated into 2 major categories – those related to research directions and those related to existing industry practices. The emerging concept of Green Software with specific focus on mainframe computing is also discussed in brief. Empirical results on power saving are included wherever available. The paper concludes that only with the close co-ordination between hardware architect, software architect and system architect low energy systems can be realized.

  9. Survey of Verification and Validation Techniques for Small Satellite Software Development

    Science.gov (United States)

    Jacklin, Stephen A.

    2015-01-01

    The purpose of this paper is to provide an overview of the current trends and practices in small-satellite software verification and validation. This document is not intended to promote a specific software assurance method. Rather, it seeks to present an unbiased survey of software assurance methods used to verify and validate small satellite software and to make mention of the benefits and value of each approach. These methods include simulation and testing, verification and validation with model-based design, formal methods, and fault-tolerant software design with run-time monitoring. Although the literature reveals that simulation and testing has by far the longest legacy, model-based design methods are proving to be useful for software verification and validation. Some work in formal methods, though not widely used for any satellites, may offer new ways to improve small satellite software verification and validation. These methods need to be further advanced to deal with the state explosion problem and to make them more usable by small-satellite software engineers to be regularly applied to software verification. Last, it is explained how run-time monitoring, combined with fault-tolerant software design methods, provides an important means to detect and correct software errors that escape the verification process or those errors that are produced after launch through the effects of ionizing radiation.

  10. The Public Health Legacy of Polio Eradication in Africa.

    Science.gov (United States)

    Craig, Allen S; Haydarov, Rustam; O'Malley, Helena; Galway, Michael; Dao, Halima; Ngongo, Ngashi; Baranyikwa, Marie Therese; Naqvi, Savita; Abid, Nima S; Pandak, Carol; Edwards, Amy

    2017-07-01

    The legacy of polio in Africa goes far beyond the tragedies of millions of children with permanent paralysis. It has a positive side, which includes the many well-trained polio staff who have vaccinated children, conducted surveillance, tested stool specimens in the laboratories, engaged with communities, and taken care of polio patients. This legacy also includes support for routine immunization services and vaccine introductions and campaigns for other diseases. As polio funding declines, it is time to take stock of the resources made available with polio funding in Africa and begin to find ways to keep some of the talented staff, infrastructure, and systems in place to work on new public health challenges. The partnerships that helped support polio eradication will need to consider funding to maintain and to strengthen routine immunization services and other maternal, neonatal, and child health programs in Africa that have benefitted from the polio eradication infrastructure. © The Author 2017. Published by Oxford University Press for the Infectious Diseases Society of America.

  11. SWEPP gamma-ray spectrometer system software user's guide

    International Nuclear Information System (INIS)

    Femec, D.A.

    1994-08-01

    The SWEPP Gamma-Ray Spectrometer (SGRS) System has been developed by the Radiation Measurement and Development Unit of the Idaho National Engineering Laboratory to assist in the characterization of the radiological contents of contact-handled waste containers at the Stored Waste Examination Pilot Plant (SWEPP). In addition to determining the concentrations of gamma-ray-emitting radionuclides, the software also calculates attenuation-corrected isotopic mass ratios of specific interest, and provides controls for SGRS hardware as required. This document serves as a user's guide for the data acquisition and analysis software associated with the SGRS system

  12. Advanced Transport Operating System (ATOPS) color displays software description: MicroVAX system

    Science.gov (United States)

    Slominski, Christopher J.; Plyler, Valerie E.; Dickson, Richard W.

    1992-01-01

    This document describes the software created for the Display MicroVAX computer used for the Advanced Transport Operating Systems (ATOPS) project on the Transport Systems Research Vehicle (TSRV). The software delivery of February 27, 1991, known as the 'baseline display system', is the one described in this document. Throughout this publication, module descriptions are presented in a standardized format which contains module purpose, calling sequence, detailed description, and global references. The global references section includes subroutines, functions, and common variables referenced by a particular module. The system described supports the Research Flight Deck (RFD) of the TSRV. The RFD contains eight Cathode Ray Tubes (CRTs) which depict a Primary Flight Display, Navigation Display, System Warning Display, Takeoff Performance Monitoring System Display, and Engine Display.

  13. Ten Years of Legacy Management: U.S. DOE Office of Legacy Management Accomplishments - 13246

    Energy Technology Data Exchange (ETDEWEB)

    Carter, Tony [U.S. Department of Energy Office of Legacy Management, 1000 Independence Ave. SW, Washington, DC 20585 (United States); Miller, Judith [S.M. Stoller Corporation, 2597 Legacy Way, Grand Junction, CO 81503 (United States)

    2013-07-01

    The U.S. Department of Energy (DOE) established the Office of Legacy Management (LM) to provide a long-term, sustainable solution to environmental impacts that remain from nuclear weapons production during World War II and the Cold War. The production activities created adverse environmental conditions at over 100 sites. When LM was established on December 15, 2003, it became responsible for 33 sites where active environmental remediation was complete. Currently, LM is responsible for long-term surveillance and maintenance of environmental remedies, promotion of beneficial reuse of land and buildings, and management of records and information at 89 sites in 29 states and Puerto Rico. LM is also responsible for meeting contractual obligations associated with former contractor workers' pensions and post-retirement benefits. Effectively addressing this environmental and human legacy will continue to require a focused and well-managed effort. (authors)

  14. Ten Years of Legacy Management: U.S. DOE Office of Legacy Management Accomplishments - 13246

    International Nuclear Information System (INIS)

    Carter, Tony; Miller, Judith

    2013-01-01

    The U.S. Department of Energy (DOE) established the Office of Legacy Management (LM) to provide a long-term, sustainable solution to environmental impacts that remain from nuclear weapons production during World War II and the Cold War. The production activities created adverse environmental conditions at over 100 sites. When LM was established on December 15, 2003, it became responsible for 33 sites where active environmental remediation was complete. Currently, LM is responsible for long-term surveillance and maintenance of environmental remedies, promotion of beneficial reuse of land and buildings, and management of records and information at 89 sites in 29 states and Puerto Rico. LM is also responsible for meeting contractual obligations associated with former contractor workers' pensions and post-retirement benefits. Effectively addressing this environmental and human legacy will continue to require a focused and well-managed effort. (authors)

  15. Honeywell Modular Automation System Computer Software Documentation

    International Nuclear Information System (INIS)

    CUNNINGHAM, L.T.

    1999-01-01

    This document provides a Computer Software Documentation for a new Honeywell Modular Automation System (MAS) being installed in the Plutonium Finishing Plant (PFP). This system will be used to control new thermal stabilization furnaces in HA-211 and vertical denitration calciner in HC-230C-2

  16. World Cup Cities Project: movement by the sport legacy of sporting mega events

    Directory of Open Access Journals (Sweden)

    Rodrigo Pojar Paiva

    2013-11-01

    Full Text Available The achievement of sporting mega events is configured as an exceptional opportunity for the promotion of sports in the host cities. For this positive legacy to happen it must be previously planned, executed and monitored. This study aimed to present the national structure of the World Cup Cities Project and analyse the results obtained in six host cities of the FIFA World Cup 2014. The results showed that the social and sports legacy was not planned by public managers and there are no systems of control and evaluation of sport in any of the host cities, until the realization of the project World Cup Cities.

  17. A Comparison and Evaluation of Real-Time Software Systems Modeling Languages

    Science.gov (United States)

    Evensen, Kenneth D.; Weiss, Kathryn Anne

    2010-01-01

    A model-driven approach to real-time software systems development enables the conceptualization of software, fostering a more thorough understanding of its often complex architecture and behavior while promoting the documentation and analysis of concerns common to real-time embedded systems such as scheduling, resource allocation, and performance. Several modeling languages have been developed to assist in the model-driven software engineering effort for real-time systems, and these languages are beginning to gain traction with practitioners throughout the aerospace industry. This paper presents a survey of several real-time software system modeling languages, namely the Architectural Analysis and Design Language (AADL), the Unified Modeling Language (UML), Systems Modeling Language (SysML), the Modeling and Analysis of Real-Time Embedded Systems (MARTE) UML profile, and the AADL for UML profile. Each language has its advantages and disadvantages, and in order to adequately describe a real-time software system's architecture, a complementary use of multiple languages is almost certainly necessary. This paper aims to explore these languages in the context of understanding the value each brings to the model-driven software engineering effort and to determine if it is feasible and practical to combine aspects of the various modeling languages to achieve more complete coverage in architectural descriptions. To this end, each language is evaluated with respect to a set of criteria such as scope, formalisms, and architectural coverage. An example is used to help illustrate the capabilities of the various languages.

  18. Software for computers in the safety systems of nuclear power stations

    International Nuclear Information System (INIS)

    1987-08-01

    This standard includes the safety actuation systems, the safety system support features and the protection systems. The standard provides requirements for each stage of software generation, including design, development, qualification and operation as well as the documentation for each stage of the software generation for the purpose of achieving highly reliable software. The principles applied in developing these requirements include: Best available practice; top-down design methods; modularity; verification of each phase; clear documentation; auditable documents and validation testing. (orig./HP)

  19. Hardware control system using modular software under RSX-11D

    International Nuclear Information System (INIS)

    Kittell, R.S.; Helland, J.A.

    1978-01-01

    A modular software system used to control extensive hardware is described. The development, operation, and experience with this software are discussed. Included are the methods employed to implement this system while taking advantage of the Real-Time features of RSX-11D. Comparisons are made between this system and an earlier nonmodular system. The controlled hardware includes magnet power supplies, stepping motors, DVM's, and multiplexors, and is interfaced through CAMAC. 4 figures

  20. Software for the ACP [Advanced Computer Program] multiprocessor system

    International Nuclear Information System (INIS)

    Biel, J.; Areti, H.; Atac, R.

    1987-01-01

    Software has been developed for use with the Fermilab Advanced Computer Program (ACP) multiprocessor system. The software was designed to make a system of a hundred independent node processors as easy to use as a single, powerful CPU. Subroutines have been developed by which a user's host program can send data to and get results from the program running in each of his ACP node processors. Utility programs make it easy to compile and link host and node programs, to debug a node program on an ACP development system, and to submit a debugged program to an ACP production system

  1. Estimating Rates of Fault Insertion and Test Effectiveness in Software Systems

    Science.gov (United States)

    Nikora, A.; Munson, J.

    1998-01-01

    In developing a software system, we would like to estimate the total number of faults inserted into a software system, the residual fault content of that system at any given time, and the efficacy of the testing activity in executing the code containing the newly inserted faults.

  2. Model-integrating software components engineering flexible software systems

    CERN Document Server

    Derakhshanmanesh, Mahdi

    2015-01-01

    In his study, Mahdi Derakhshanmanesh builds on the state of the art in modeling by proposing to integrate models into running software on the component-level without translating them to code. Such so-called model-integrating software exploits all advantages of models: models implicitly support a good separation of concerns, they are self-documenting and thus improve understandability and maintainability and in contrast to model-driven approaches there is no synchronization problem anymore between the models and the code generated from them. Using model-integrating components, software will be

  3. The mycological legacy of Elias Magnus Fries

    DEFF Research Database (Denmark)

    Petersen, Ronald H.; Knudsen, Henning

    2015-01-01

    : students and associates aided Fries and after his passing carried forward his taxonomic ideas. His legacy spawned a line of Swedish and Danish mycologists intent on perpetuating the Fries tradition: Hampus von Post, Lars Romell, Seth Lundell and John Axel Nannfeldt in Sweden; Emil Rostrup, Severin Petersen...... with clarity. In the 20th century, nomenclatural commissions legislated Fries's Systema and Elenchus as the "starting point" for names of most fungi, giving these books special recognition. The present paper attempts to trace Fries's legacy from his lifetime to the recent past....

  4. The Software Bus, an Object-Oriented Data Exchange System

    International Nuclear Information System (INIS)

    Akerbaek, T.; Louka, M.

    1996-01-01

    This document describes the Software Bus System, developed for object-oriented task to task communication in a TCP/IP based network. The Software Bus is a set of library functions, developed to be used for the Picasso-3 UIMS, and as a general purpose tool for dynamically interfacing programs at run-time. The Software Bus offers a high level object-oriented data exchange mechanism that relieves the application programmer of the low level TCP/IP-programming and communication protocol handling. The Software Bus is currently available under several UNIX platforms and a version for Windows NT is planned for late 1996. (author)

  5. Jack Wescott and Donald F. Smith. The Legacy Project

    Science.gov (United States)

    Moye, Johnny J.; Wescott, Jack W.; Smith, Donald F.

    2017-01-01

    This is the tenth in a series of articles entitled "The Legacy Project." The Legacy Project focuses on the lives and actions of leaders who have forged our profession into what it is today. Members of the profession owe a debt of gratitude to these leaders. One simple way to demonstrate that gratitude is to recognize these leaders and…

  6. Software management of the LHC Detector Control Systems

    CERN Document Server

    Varela, F

    2007-01-01

    The control systems of each of the four Large Hadron Collider (LHC) experiments will contain of the order of 150 computers running the back-end applications. These applications will have to be maintained and eventually upgraded during the lifetime of the experiments, ~20 years. This paper presents the centralized software management strategy adopted by the Joint COntrols Project (JCOP) [1], which is based on a central database that holds the overall system configuration. The approach facilitates the integration of different parts of a control system and provides versioning of its various software components. The information stored in the configuration database can eventually be used to restore a computer in the event of failure.

  7. Software management of the LHC detector control systems

    CERN Document Server

    Varela, F

    2007-01-01

    The control systems of each of the four Large Hadron Collider (LHC) experiments will contain of the order of 150 computers running the back-end applications. These applications will have to be maintained and eventually upgraded during the lifetime of the experiments, ~20 years. This paper presents the centralized software management strategy adopted by the Joint COntrols Project (JCOP) [1], which is based on a central database that holds the overall system configuration. The approach facilitates the integration of different parts of a control system and provides versioning of its various software components. The information stored in the configuration database can eventually be used to restore a computer in the event of failure.

  8. Development of a support software system for real-time HAL/S applications

    Science.gov (United States)

    Smith, R. S.

    1984-01-01

    Methodologies employed in defining and implementing a software support system for the HAL/S computer language for real-time operations on the Shuttle are detailed. Attention is also given to the management and validation techniques used during software development and software maintenance. Utilities developed to support the real-time operating conditions are described. With the support system being produced on Cyber computers and executable code then processed through Cyber or PDP machines, the support system has a production level status and can serve as a model for other software development projects.

  9. Development of design and analysis software for advanced nuclear system

    International Nuclear Information System (INIS)

    Wu Yican; Hu Liqin; Long Pengcheng; Luo Yuetong; Li Yazhou; Zeng Qin; Lu Lei; Zhang Junjun; Zou Jun; Xu Dezheng; Bai Yunqing; Zhou Tao; Chen Hongli; Peng Lei; Song Yong; Huang Qunying

    2010-01-01

    A series of professional codes, which are necessary software tools and data libraries for advanced nuclear system design and analysis, were developed by the FDS Team, including the codes of automatic modeling, physics and engineering calculation, virtual simulation and visualization, system engineering and safety analysis and the related database management etc. The development of these software series was proposed as an exercise of development of nuclear informatics. This paper introduced the main functions and key techniques of the software series, as well as some tests and practical applications. (authors)

  10. AWARE-P: a collaborative, system-based IAM planning software

    OpenAIRE

    Coelho, S. T.; Vitorino, D.

    2011-01-01

    The AWARE-P project aims to promote the application of integrated and risk-based approaches to the rehabilitation of urban water supply and wastewater drainage systems. Central to the project is the development of a software platform based on a set of computational components, which assist in the analyses and decision support involved in the planning process for sustainable infrastructural asset management. The AWARE-P software system brings together onto a common platform the inf...

  11. Scientific Data as the Core Legacy of IPY

    Science.gov (United States)

    Parsons, M. A.

    2008-12-01

    The interdisciplinary breadth of the International Polar Year is unprecedented. The IPY has explicit objectives to link researchers across different fields to address questions and issues lying beyond the scope of individual disciplines and to strengthen international coordination of research and enhance international collaboration and cooperation. The IPY Data Policy and Management Subcommittee have developed a policy to help meet these objectives and an international collaboration of investigators and data managers, the IPY Data and Information Service, are working to make IPY data widely available. I will present an overview of the primary data management considerations for IPY and how diverse organizations are making IPY and related data available. Centralized discovery mechanisms for widely distributed data plus targeted access mechanisms for specific disciplines will be presented. These range from near real time access to satellite remote sensing data and GCM output to fair and appropriate access to traditional knowledge of the Arctic. These mechanisms reflect significant advancement in polar data management, but they belie the major challenges that remain. These challenges include fostering a culture change in science that puts greater value on data publication and open data access as well as developing sustained systems and business models for the long-term preservation of IPY data. This will be crucial to ensuring the legacy of IPY, a major objective of IPY sponsors, ICSU and WMO. New efforts to ensure this legacy include the development of the WMO Information System, the Sustained Arctic Observing Network, and the Global Earth Observing System of Systems; the reform of ICSU's World Data Center System; and the results of the Electronic Geophysical Year.

  12. Error detection and prevention in Embedded Systems Software

    DEFF Research Database (Denmark)

    Kamel, Hani Fouad

    1996-01-01

    Despite many efforts to structure the development and design processes of embedded systems, errors are discovered at the final stages of production and sometimes after the delivery of the products. The cost of such errors can be prohibitive.Different design techniques to detect such errors...... systems, a formal model for such systems is introduced. The main characteristics of embedded systems design and the interaction of these properties are described. A taxonomy for the structure of the software developed for such systems based on the amount of processes and processors involved is presented.......The second part includes methods and techniques to detect software design errors.The third part deals with error prevention. It starts with a presentation of different models of the development processes used in industry and taught at universities. This leads us to deduce the major causes of errors...

  13. Software application for quality control protocol of mammography systems

    International Nuclear Information System (INIS)

    Kjosevski, Vladimir; Gershan, Vesna; Ginovska, Margarita; Spasevska, Hristina

    2010-01-01

    Considering the fact that the Quality Control of the technological process of the mammographic system involves testing of a large number of parameters, it is clearly evident that there is a need for using the information technology for gathering, processing and storing of all the parameters that are result of this process. The main goal of this software application is facilitation and automation of the gathering, processing, storing and presenting process of the data related to the qualification of the physical and technical parameters during the quality control of the technological process of the mammographic system. The software application along with its user interface and database has been made with the Microsoft Access 2003 application which is part of the Microsoft Office 2003 software packet and has been chosen as a platform for developing because it is the most commonly used office application today among the computer users in the country. This is important because it will provide the end users a familiar environment to work in, without the need for additional training and improving the computer skills that they posses. Most importantly, the software application is easy to use, fast in calculating the parameters needed and it is an excellent way to store and display the results. There is a possibility for up scaling this software solution so it can be used by many different users at the same time over the Internet. It is highly recommended that this system is implemented as soon as possible in the quality control process of the mammographic systems due to its many advantages.(Author)

  14. Sports Mega-Events: Reflections About the Legacies of Uefa Euro Cup

    Directory of Open Access Journals (Sweden)

    Fillipe Soares Romano

    2016-01-01

    Full Text Available This essay conducted a literature review of the legacy of sports mega-events, emphasizing the UEFA European Championship. In order to review the academic literature about the subject, we used the following keywords: legacies; sports mega-events and its legacies; UEFA European Championship; and their corresponding terms in Spanish and Portuguese. It was also reviewed some important research basis: Annals; BVS; DOAJ; EBSCO; REDALYC; Periódicos CAPES; PubMed; Scielo; Scopus; Science Direct; and, Scholar Google. The results were analyzed by categories. Regarding the sport mega-event Eurocopa, it was found that the tourism legacies, the economics impacts and the social aspect have been discussed more often, as well as projections and predictions of events that will occur later. It was also concluded that most of the articles shows an absence of discussion about the negative impacts of sports mega-events. It is therefore recommended to use standard methods with comparisons of results and longitudinal studies to better understand the impacts and the use of legacy.

  15. Considerations of the Software Metric-based Methodology for Software Reliability Assessment in Digital I and C Systems

    International Nuclear Information System (INIS)

    Ha, J. H.; Kim, M. K.; Chung, B. S.; Oh, H. C.; Seo, M. R.

    2007-01-01

    Analog I and C systems have been replaced by digital I and C systems because the digital systems have many potential benefits to nuclear power plants in terms of operational and safety performance. For example, digital systems are essentially free of drifts, have higher data handling and storage capabilities, and provide improved performance by accuracy and computational capabilities. In addition, analog replacement parts become more difficult to obtain since they are obsolete and discontinued. There are, however, challenges to the introduction of digital technology into the nuclear power plants because digital systems are more complex than analog systems and their operation and failure modes are different. Especially, software, which can be the core of functionality in the digital systems, does not wear out physically like hardware and its failure modes are not yet defined clearly. Thus, some researches to develop the methodology for software reliability assessment are still proceeding in the safety-critical areas such as nuclear system, aerospace and medical devices. Among them, software metric-based methodology has been considered for the digital I and C systems of Korean nuclear power plants. Advantages and limitations of that methodology are identified and requirements for its application to the digital I and C systems are considered in this study

  16. Till Moritz Karbach, Scientific Legacy

    CERN Document Server

    Aaij, Roel; Adinolfi, Marco; Affolder, Anthony; Ajaltouni, Ziad; Akar, Simon; Albrecht, Johannes; Alessio, Federico; Alexander, Michael; Ali, Suvayu; Alkhazov, Georgy; Alvarez Cartelle, Paula; Alves Jr, Antonio Augusto; Amato, Sandra; Amerio, Silvia; Amhis, Yasmine; An, Liupan; Anderlini, Lucio; Anderson, Jonathan; Andreassi, Guido; Andreotti, Mirco; Andrews, Jason; Appleby, Robert; Aquines Gutierrez, Osvaldo; Archilli, Flavio; d'Argent, Philippe; Artamonov, Alexander; Artuso, Marina; Aslanides, Elie; Auriemma, Giulio; Baalouch, Marouen; Bachmann, Sebastian; Back, John; Badalov, Alexey; Baesso, Clarissa; Baldini, Wander; Barlow, Roger; Barschel, Colin; Barsuk, Sergey; Barter, William; Batozskaya, Varvara; Battista, Vincenzo; Bay, Aurelio; Beaucourt, Leo; Beddow, John; Bedeschi, Franco; Bediaga, Ignacio; Bel, Lennaert; Bellee, Violaine; Belyaev, Ivan; Ben-Haim, Eli; Bencivenni, Giovanni; Benson, Sean; Benton, Jack; Berezhnoy, Alexander; Bernet, Roland; Bertolin, Alessandro; Bettler, Marc-Olivier; van Beuzekom, Martinus; Bien, Alexander; Bifani, Simone; Bird, Thomas; Birnkraut, Alex; Bizzeti, Andrea; Blake, Thomas; Blanc, Frédéric; Blouw, Johan; Blusk, Steven; Bocci, Valerio; Bondar, Alexander; Bondar, Nikolay; Bonivento, Walter; Borghi, Silvia; Borsato, Martino; Bowcock, Themistocles; Bowen, Espen Eie; Bozzi, Concezio; Braun, Svende; Brett, David; Britsch, Markward; Britton, Thomas; Brodzicka, Jolanta; Brook, Nicholas; Bursche, Albert; Buytaert, Jan; Cadeddu, Sandro; Calabrese, Roberto; Calvi, Marta; Calvo Gomez, Miriam; Campana, Pierluigi; Campora Perez, Daniel; Capriotti, Lorenzo; Carbone, Angelo; Carboni, Giovanni; Cardinale, Roberta; Cardini, Alessandro; Carniti, Paolo; Carson, Laurence; Carvalho Akiba, Kazuyoshi; Casse, Gianluigi; Cassina, Lorenzo; Castillo Garcia, Lucia; Cattaneo, Marco; Cauet, Christophe; Cavallero, Giovanni; Cenci, Riccardo; Charles, Matthew; Charpentier, Philippe; Chefdeville, Maximilien; Chen, Shanzhen; Cheung, Shu-Faye; Chiapolini, Nicola; Chrzaszcz, Marcin; Cid Vidal, Xabier; Ciezarek, Gregory; Clarke, Peter; Clemencic, Marco; Cliff, Harry; Closier, Joel; Coco, Victor; Cogan, Julien; Cogneras, Eric; Cogoni, Violetta; Cojocariu, Lucian; Collazuol, Gianmaria; Collins, Paula; Comerma-Montells, Albert; Contu, Andrea; Cook, Andrew; Coombes, Matthew; Coquereau, Samuel; Corti, Gloria; Corvo, Marco; Couturier, Benjamin; Cowan, Greig; Craik, Daniel Charles; Crocombe, Andrew; Cruz Torres, Melissa Maria; Cunliffe, Samuel; Currie, Robert; D'Ambrosio, Carmelo; Dalseno, Jeremy; David, Pieter; Davis, Adam; De Bruyn, Kristof; De Capua, Stefano; De Cian, Michel; De Miranda, Jussara; De Paula, Leandro; De Silva, Weeraddana; De Simone, Patrizia; Dean, Cameron Thomas; Decamp, Daniel; Deckenhoff, Mirko; Del Buono, Luigi; Déléage, Nicolas; Demmer, Moritz; Derkach, Denis; Deschamps, Olivier; Dettori, Francesco; Dey, Biplab; Di Canto, Angelo; Di Ruscio, Francesco; Dijkstra, Hans; Donleavy, Stephanie; Dordei, Francesca; Dorigo, Mirco; Dosil Suárez, Alvaro; Dossett, David; Dovbnya, Anatoliy; Dreimanis, Karlis; Dufour, Laurent; Dujany, Giulio; Dupertuis, Frederic; Durante, Paolo; Dzhelyadin, Rustem; Dziurda, Agnieszka; Dzyuba, Alexey; Easo, Sajan; Egede, Ulrik; Egorychev, Victor; Eidelman, Semen; Eisenhardt, Stephan; Eitschberger, Ulrich; Ekelhof, Robert; Eklund, Lars; El Rifai, Ibrahim; Elsasser, Christian; Ely, Scott; Esen, Sevda; Evans, Hannah Mary; Evans, Timothy; Falabella, Antonio; Färber, Christian; Farinelli, Chiara; Farley, Nathanael; Farry, Stephen; Fay, Robert; Ferguson, Dianne; Fernandez Albor, Victor; Ferrari, Fabio; Ferreira Rodrigues, Fernando; Ferro-Luzzi, Massimiliano; Filippov, Sergey; Fiore, Marco; Fiorini, Massimiliano; Firlej, Miroslaw; Fitzpatrick, Conor; Fiutowski, Tomasz; Fohl, Klaus; Fol, Philip; Fontana, Marianna; Fontanelli, Flavio; Forty, Roger; Francisco, Oscar; Frank, Markus; Frei, Christoph; Frosini, Maddalena; Fu, Jinlin; Furfaro, Emiliano; Gallas Torreira, Abraham; Galli, Domenico; Gallorini, Stefano; Gambetta, Silvia; Gandelman, Miriam; Gandini, Paolo; Gao, Yuanning; García Pardiñas, Julián; Garra Tico, Jordi; Garrido, Lluis; Gascon, David; Gaspar, Clara; Gastaldi, Ugo; Gauld, Rhorry; Gavardi, Laura; Gazzoni, Giulio; Geraci, Angelo; Gerick, David; Gersabeck, Evelina; Gersabeck, Marco; Gershon, Timothy; Ghez, Philippe; Gianelle, Alessio; Gianì, Sebastiana; Gibson, Valerie; Girard, Olivier Göran; Giubega, Lavinia-Helena; Gligorov, Vladimir; Göbel, Carla; Golubkov, Dmitry; Golutvin, Andrey; Gomes, Alvaro; Gotti, Claudio; Grabalosa Gándara, Marc; Graciani Diaz, Ricardo; Granado Cardoso, Luis Alberto; Graugés, Eugeni; Graverini, Elena; Graziani, Giacomo; Grecu, Alexandru; Greening, Edward; Gregson, Sam; Griffith, Peter; Grillo, Lucia; Grünberg, Oliver; Gui, Bin; Gushchin, Evgeny; Guz, Yury; Gys, Thierry; Hadavizadeh, Thomas; Hadjivasiliou, Christos; Haefeli, Guido; Haen, Christophe; Haines, Susan; Hall, Samuel; Hamilton, Brian; Han, Xiaoxue; Hansmann-Menzemer, Stephanie; Harnew, Neville; Harnew, Samuel; Harrison, Jonathan; He, Jibo; Head, Timothy; Heijne, Veerle; Hennessy, Karol; Henrard, Pierre; Henry, Louis; Hernando Morata, Jose Angel; van Herwijnen, Eric; Heß, Miriam; Hicheur, Adlène; Hill, Donal; Hoballah, Mostafa; Hombach, Christoph; Hulsbergen, Wouter; Humair, Thibaud; Hussain, Nazim; Hutchcroft, David; Hynds, Daniel; Idzik, Marek; Ilten, Philip; Jacobsson, Richard; Jaeger, Andreas; Jalocha, Pawel; Jans, Eddy; Jawahery, Abolhassan; Jing, Fanfan; John, Malcolm; Johnson, Daniel; Jones, Christopher; Joram, Christian; Jost, Beat; Jurik, Nathan; Kandybei, Sergii; Kanso, Walaa; Karacson, Matthias; Karodia, Sarah; Kelsey, Matthew; Kenyon, Ian; Kenzie, Matthew; Ketel, Tjeerd; Khanji, Basem; Khurewathanakul, Chitsanu; Klaver, Suzanne; Klimaszewski, Konrad; Kochebina, Olga; Kolpin, Michael; Komarov, Ilya; Koopman, Rose; Koppenburg, Patrick; Korolev, Mikhail; Kozeiha, Mohamad; Kravchuk, Leonid; Kreplin, Katharina; Kreps, Michal; Krocker, Georg; Krokovny, Pavel; Kruse, Florian; Kucewicz, Wojciech; Kucharczyk, Marcin; Kudryavtsev, Vasily; Kuonen, Axel Kevin; Kurek, Krzysztof; Kvaratskheliya, Tengiz; Lacarrere, Daniel; Lafferty, George; Lai, Adriano; Lambert, Dean; Lambert, Robert W; Lanfranchi, Gaia; Langenbruch, Christoph; Langhans, Benedikt; Latham, Thomas; Lazzeroni, Cristina; Le Gac, Renaud; van Leerdam, Jeroen; Lees, Jean-Pierre; Lefèvre, Regis; Leflat, Alexander; Lefrançois, Jacques; Leroy, Olivier; Lesiak, Tadeusz; Leverington, Blake; Li, Yiming; Likhomanenko, Tatiana; Liles, Myfanwy; Lindner, Rolf; Linn, Christian; Lionetto, Federica; Liu, Bo; Liu, Xuesong; Loh, David; Lohn, Stefan; Longstaff, Iain; Lopes, Jose; Lowdon, Peter; Lucchesi, Donatella; Luo, Haofei; Lupato, Anna; Luppi, Eleonora; Lupton, Oliver; Machefert, Frederic; Maciuc, Florin; Maev, Oleg; Maguire, Kevin; Malde, Sneha; Malinin, Alexander; Manca, Giulia; Mancinelli, Giampiero; Manning, Peter Michael; Mapelli, Alessandro; Maratas, Jan; Marchand, Jean François; Marconi, Umberto; Marin Benito, Carla; Marino, Pietro; Märki, Raphael; Marks, Jörg; Martellotti, Giuseppe; Martin, Morgan; Martinelli, Maurizio; Martinez Santos, Diego; Martinez Vidal, Fernando; Martins Tostes, Danielle; Massafferri, André; Matev, Rosen; Mathad, Abhijit; Mathe, Zoltan; Matteuzzi, Clara; Mauri, Andrea; Maurin, Brice; Mazurov, Alexander; McCann, Michael; McCarthy, James; McNab, Andrew; McNulty, Ronan; Meadows, Brian; Meier, Frank; Meissner, Marco; Melnychuk, Dmytro; Merk, Marcel; Milanes, Diego Alejandro; Minard, Marie-Noelle; Mitzel, Dominik Stefan; Molina Rodriguez, Josue; Monteil, Stephane; Morandin, Mauro; Morawski, Piotr; Mordà, Alessandro; Morello, Michael Joseph; Moron, Jakub; Morris, Adam Benjamin; Mountain, Raymond; Muheim, Franz; Müller, Janine; Müller, Katharina; Müller, Vanessa; Mussini, Manuel; Muster, Bastien; Naik, Paras; Nakada, Tatsuya; Nandakumar, Raja; Nandi, Anita; Nasteva, Irina; Needham, Matthew; Neri, Nicola; Neubert, Sebastian; Neufeld, Niko; Neuner, Max; Nguyen, Anh Duc; Nguyen, Thi-Dung; Nguyen-Mau, Chung; Niess, Valentin; Niet, Ramon; Nikitin, Nikolay; Nikodem, Thomas; Ninci, Daniele; Novoselov, Alexey; O'Hanlon, Daniel Patrick; Oblakowska-Mucha, Agnieszka; Obraztsov, Vladimir; Ogilvy, Stephen; Okhrimenko, Oleksandr; Oldeman, Rudolf; Onderwater, Gerco; Osorio Rodrigues, Bruno; Otalora Goicochea, Juan Martin; Otto, Adam; Owen, Patrick; Oyanguren, Maria Aranzazu; Palano, Antimo; Palombo, Fernando; Palutan, Matteo; Panman, Jacob; Papanestis, Antonios; Pappagallo, Marco; Pappalardo, Luciano; Pappenheimer, Cheryl; Parkes, Christopher; Passaleva, Giovanni; Patel, Girish; Patel, Mitesh; Patrignani, Claudia; Pearce, Alex; Pellegrino, Antonio; Penso, Gianni; Pepe Altarelli, Monica; Perazzini, Stefano; Perret, Pascal; Pescatore, Luca; Petridis, Konstantinos; Petrolini, Alessandro; Petruzzo, Marco; Picatoste Olloqui, Eduardo; Pietrzyk, Boleslaw; Pilař, Tomas; Pinci, Davide; Pistone, Alessandro; Piucci, Alessio; Playfer, Stephen; Plo Casasus, Maximo; Poikela, Tuomas; Polci, Francesco; Poluektov, Anton; Polyakov, Ivan; Polycarpo, Erica; Popov, Alexander; Popov, Dmitry; Popovici, Bogdan; Potterat, Cédric; Price, Eugenia; Price, Joseph David; Prisciandaro, Jessica; Pritchard, Adrian; Prouve, Claire; Pugatch, Valery; Puig Navarro, Albert; Punzi, Giovanni; Qian, Wenbin; Quagliani, Renato; Rachwal, Bartolomiej; Rademacker, Jonas; Rama, Matteo; Rangel, Murilo; Raniuk, Iurii; Rauschmayr, Nathalie; Raven, Gerhard; Redi, Federico; Reichert, Stefanie; Reid, Matthew; dos Reis, Alberto; Ricciardi, Stefania; Richards, Sophie; Rihl, Mariana; Rinnert, Kurt; Rives Molina, Vincente; Robbe, Patrick; Rodrigues, Ana Barbara; Rodrigues, Eduardo; Rodriguez Lopez, Jairo Alexis; Rodriguez Perez, Pablo; Roiser, Stefan; Romanovsky, Vladimir; Romero Vidal, Antonio; Ronayne, John William; Rotondo, Marcello; Rouvinet, Julien; Ruf, Thomas; Ruiz, Hugo; Ruiz Valls, Pablo; Saborido Silva, Juan Jose; Sagidova, Naylya; Sail, Paul; Saitta, Biagio; Salustino Guimaraes, Valdir; Sanchez Mayordomo, Carlos; Sanmartin Sedes, Brais; Santacesaria, Roberta; Santamarina Rios, Cibran; Santimaria, Marco; Santovetti, Emanuele; Sarti, Alessio; Satriano, Celestina; Satta, Alessia; Saunders, Daniel Martin; Savrina, Darya; Schiller, Manuel; Schindler, Heinrich; Schlupp, Maximilian; Schmelling, Michael; Schmelzer, Timon; Schmidt, Burkhard; Schneider, Olivier; Schopper, Andreas; Schubiger, Maxime; Schune, Marie Helene; Schwemmer, Rainer; Sciascia, Barbara; Sciubba, Adalberto; Semennikov, Alexander; Serra, Nicola; Serrano, Justine; Sestini, Lorenzo; Seyfert, Paul; Shapkin, Mikhail; Shapoval, Illya; Shcheglov, Yury; Shears, Tara; Shekhtman, Lev; Shevchenko, Vladimir; Shires, Alexander; Silva Coutinho, Rafael; Simi, Gabriele; Sirendi, Marek; Skidmore, Nicola; Skillicorn, Ian; Skwarnicki, Tomasz; Smith, Edmund; Smith, Eluned; Smith, Iwan Thomas; Smith, Jackson; Smith, Mark; Snoek, Hella; Sokoloff, Michael; Soler, Paul; Soomro, Fatima; Souza, Daniel; Souza De Paula, Bruno; Spaan, Bernhard; Spradlin, Patrick; Sridharan, Srikanth; Stagni, Federico; Stahl, Marian; Stahl, Sascha; Steinkamp, Olaf; Stenyakin, Oleg; Sterpka, Christopher Francis; Stevenson, Scott; Stoica, Sabin; Stone, Sheldon; Storaci, Barbara; Stracka, Simone; Straticiuc, Mihai; Straumann, Ulrich; Sun, Liang; Sutcliffe, William; Swientek, Krzysztof; Swientek, Stefan; Syropoulos, Vasileios; Szczekowski, Marek; Szczypka, Paul; Szumlak, Tomasz; T'Jampens, Stephane; Tekampe, Tobias; Teklishyn, Maksym; Tellarini, Giulia; Teubert, Frederic; Thomas, Christopher; Thomas, Eric; van Tilburg, Jeroen; Tisserand, Vincent; Tobin, Mark; Todd, Jacob; Tolk, Siim; Tomassetti, Luca; Tonelli, Diego; Topp-Joergensen, Stig; Torr, Nicholas; Tournefier, Edwige; Tourneur, Stephane; Trabelsi, Karim; Tran, Minh Tâm; Tresch, Marco; Trisovic, Ana; Tsaregorodtsev, Andrei; Tsopelas, Panagiotis; Tuning, Niels; Ukleja, Artur; Ustyuzhanin, Andrey; Uwer, Ulrich; Vacca, Claudia; Vagnoni, Vincenzo; Valenti, Giovanni; Vallier, Alexis; Vazquez Gomez, Ricardo; Vazquez Regueiro, Pablo; Vázquez Sierra, Carlos; Vecchi, Stefania; Velthuis, Jaap; Veltri, Michele; Veneziano, Giovanni; Vesterinen, Mika; Viaud, Benoit; Vieira, Daniel; Vieites Diaz, Maria; Vilasis-Cardona, Xavier; Vollhardt, Achim; Volyanskyy, Dmytro; Voong, David; Vorobyev, Alexey; Vorobyev, Vitaly; Voß, Christian; de Vries, Jacco; Waldi, Roland; Wallace, Charlotte; Wallace, Ronan; Walsh, John; Wandernoth, Sebastian; Wang, Jianchun; Ward, David; Watson, Nigel; Websdale, David; Weiden, Andreas; Whitehead, Mark; Wiedner, Dirk; Wilkinson, Guy; Wilkinson, Michael; Williams, Mark Richard James; Williams, Matthew; Williams, Mike; Williams, Timothy; Wilson, Fergus; Wimberley, Jack; Wishahi, Julian; Wislicki, Wojciech; Witek, Mariusz; Wormser, Guy; Wotton, Stephen; Wright, Simon; Wyllie, Kenneth; Xie, Yuehong; Xu, Zhirui; Yang, Zhenwei; Yu, Jiesheng; Yuan, Xuhao; Yushchenko, Oleg; Zangoli, Maria; Zavertyaev, Mikhail; Zhang, Liming; Zhang, Yanxi; Zhelezov, Alexey; Zhokhov, Anatoly; Zhong, Liang

    2015-01-01

    We are deeply touched by the sudden loss of our dear friend and colleague Till Moritz Karbach. With this memorial book we wish to commemorate Moritz’ scientific legacy, and what Moritz meant to us as a friend.

  17. Software tools for microprocessor based systems

    International Nuclear Information System (INIS)

    Halatsis, C.

    1981-01-01

    After a short review of the hardware and/or software tools for the development of single-chip, fixed instruction set microprocessor-based sytems we focus on the software tools for designing systems based on microprogrammed bit-sliced microprocessors. Emphasis is placed on meta-microassemblers and simulation facilties at the register-transfer-level and architecture level. We review available meta-microassemblers giving their most important features, advantages and disadvantages. We also make extentions to higher-level microprogramming languages and associated systems specifically developed for bit-slices. In the area of simulation facilities we first discuss the simulation objectives and the criteria for chosing the right simulation language. We consertrate to simulation facilities already used in bit-slices projects and discuss the gained experience. We conclude by describing the way the Signetics meta-microassembler and the ISPS simulation tool have been employed in the design of a fast microprogrammed machine, called MICE, made out of ECL bit-slices. (orig.)

  18. Software for computers in safety systems of nuclear power plants

    International Nuclear Information System (INIS)

    Gallagher, J.M.

    1983-01-01

    The application of distributed digital processing techniques to the protection systems of nuclear power plants provides a means to significantly improve the functional capability of the protection system with respect to the operability and availability of the power plant. A major factor in the realization of this improvement is the development and maintenance of essentially error-free software. A joint program for the development of principles for the design, testing and documentation of software to achieve this goal is presented. Results from two separate experiences in the application of these principles in terms of detected software errors are summarized. The low number of errors detected during the verification testing phase demonstrates the effectiveness of the design and documentation principles in the realization of highly reliable software. (author)

  19. Timing system control software in the SLC

    International Nuclear Information System (INIS)

    Thompson, K.; Phinney, N.

    1985-04-01

    A new timing system that allows precision (approx.1 to 2 ns) control of the trigger times of klystrons, beam position monitors, and other devices on a pulse-to-pulse basis at up to 360 Hz is in operation in the first third of the SLAC linear accelerator. The control software is divided between a central host VAX and local Intel 8086-based microprocessor clusters. Facilities exist to set up and adjust the timing of devices or groups of devices independently for beam pulses having different destinations and purposes, which are run in an interlaced fashion during normal machine operation. Upgrading of the system is currently underway, using a new version of the Programmable Delay Unit CAMAC module to allow pipelining of timing information for three machine pulses. An overview of the current state of the system is presented in this paper, with an emphasis on software control

  20. Software cost/resource modeling: Software quality tradeoff measurement

    Science.gov (United States)

    Lawler, R. W.

    1980-01-01

    A conceptual framework for treating software quality from a total system perspective is developed. Examples are given to show how system quality objectives may be allocated to hardware and software; to illustrate trades among quality factors, both hardware and software, to achieve system performance objectives; and to illustrate the impact of certain design choices on software functionality.

  1. Župančič's Legacy and His Personal Library

    Directory of Open Access Journals (Sweden)

    Dragica Trobec Zadnik

    2015-12-01

    Full Text Available The legacy of writer Oton Župančič, acquired by the City Museum of Ljubljana in 1985 with a donation of the artist’s family, is a rare so large and diverse among artistic legacies. Its key part is the furniture of two rooms from Župančič’s last homes with his personal library. This article presents an overview of museum work done, from the acquisition of the legacy to its first public presentation in the framework of the exhibition Župančičeva spominska zbirka in 1985 in the museum. The methodology of the inventarization and technical foundation for the second exhibition of Oton Župančič legacy in 2008 is described. It is one of the few museum exhibitions facilitating the lovers of literary art to be in touch with the artist’s legacy on a daily basis. More detailed description of the museum’s inventory of the personal library uncovered that books are not just a literal material but also the museum subjects that exhibit artist’s work and interests and as such are the mirror of time. Many books overwritten and painted became the artist’s notebooks and thus his personal belongings. In conclusion, the obtained results and limitations are discussed and plans for further consideration and presentation of the legacy are mentioned.

  2. Instrument control software development process for the multi-star AO system ARGOS

    Science.gov (United States)

    Kulas, M.; Barl, L.; Borelli, J. L.; Gässler, W.; Rabien, S.

    2012-09-01

    The ARGOS project (Advanced Rayleigh guided Ground layer adaptive Optics System) will upgrade the Large Binocular Telescope (LBT) with an AO System consisting of six Rayleigh laser guide stars. This adaptive optics system integrates several control loops and many different components like lasers, calibration swing arms and slope computers that are dispersed throughout the telescope. The purpose of the instrument control software (ICS) is running this AO system and providing convenient client interfaces to the instruments and the control loops. The challenges for the ARGOS ICS are the development of a distributed and safety-critical software system with no defects in a short time, the creation of huge and complex software programs with a maintainable code base, the delivery of software components with the desired functionality and the support of geographically distributed project partners. To tackle these difficult tasks, the ARGOS software engineers reuse existing software like the novel middleware from LINC-NIRVANA, an instrument for the LBT, provide many tests at different functional levels like unit tests and regression tests, agree about code and architecture style and deliver software incrementally while closely collaborating with the project partners. Many ARGOS ICS components are already successfully in use in the laboratories for testing ARGOS control loops.

  3. Quality assurance of the modernized Dukovany I and C safety system software

    International Nuclear Information System (INIS)

    Karpeta, C.

    2005-01-01

    The approach to quality assurance of the software that implements the instrumentation and control functions for safety category A as per IEC 61226, which has been adopted within the 'NPP Dukovany I and C Refurbishment' project, is described. A survey of the requirements for software quality assurance of the systems that initiate protection interventions in the event of anticipated operational occurrences or accident conditions is given. The software development process applied by the system designers and manufacturers, from the software requirements specification phase to the software testing phase, is outlined. Basic information on technical audits of the software development process is also provided. (orig.)

  4. Implicit Particle Filter for Power System State Estimation with Large Scale Renewable Power Integration.

    Science.gov (United States)

    Uzunoglu, B.; Hussaini, Y.

    2017-12-01

    Implicit Particle Filter is a sequential Monte Carlo method for data assimilation that guides the particles to the high-probability by an implicit step . It optimizes a nonlinear cost function which can be inherited from legacy assimilation routines . Dynamic state estimation for almost real-time applications in power systems are becomingly increasingly more important with integration of variable wind and solar power generation. New advanced state estimation tools that will replace the old generation state estimation in addition to having a general framework of complexities should be able to address the legacy software and able to integrate the old software in a mathematical framework while allowing the power industry need for a cautious and evolutionary change in comparison to a complete revolutionary approach while addressing nonlinearity and non-normal behaviour. This work implements implicit particle filter as a state estimation tool for the estimation of the states of a power system and presents the first implicit particle filter application study on a power system state estimation. The implicit particle filter is introduced into power systems and the simulations are presented for a three-node benchmark power system . The performance of the filter on the presented problem is analyzed and the results are presented.

  5. IndElec: A Software for Analyzing Party Systems and Electoral Systems

    Directory of Open Access Journals (Sweden)

    Francisco Ocaña

    2011-08-01

    Full Text Available IndElec is a software addressed to compute a wide range of indices from electoral data, which are intended to analyze both party systems and electoral systems in political studies. Further, IndElec can calculate such indices from electoral data at several levels of aggregation, even when the acronyms of some political parties change across districts. As the amount of information provided by IndElec may be considerable, this software also aids the user in the analysis of electoral data through three capabilities. First, IndElec automatically elaborates preliminary descriptive statistical reports of computed indices. Second, IndElec saves the computed information into text files in data matrix format, which can be directly loaded by any statistical software to facilitate more sophisticated statistical studies. Third, IndElec provides results in several file formats (text, CSV, HTML, R to facilitate their visualization and management by using a wide range of application softwares (word processors, spreadsheets, web browsers, etc.. Finally, a graphical user interface is provided for IndElec to manage calculation processes, but no visualization facility is available in this environment. In fact, both the inputs and outputs for IndElec are arranged in files with the aforementioned formats.

  6. Featureous: A Tool for Feature-Centric Analysis of Java Software

    DEFF Research Database (Denmark)

    Olszak, Andrzej; Jørgensen, Bo Nørregaard

    2010-01-01

    Feature-centric comprehension of source code is necessary for incorporating user-requested modifications during software evolution and maintenance. However, such comprehension is difficult to achieve in case of large object-oriented programs due to the size, complexity, and implicit character...... of mappings between features and source code. To support programmers in overcoming these difficulties, we present a feature-centric analysis tool, Featureous. Our tool extends the NetBeans IDE with mechanisms for efficient location of feature implementations in legacy source code, and an extensive analysis...

  7. ISWHM: Tools and Techniques for Software and System Health Management

    Science.gov (United States)

    Schumann, Johann; Mengshoel, Ole J.; Darwiche, Adnan

    2010-01-01

    This presentation presents status and results of research on Software Health Management done within the NRA "ISWHM: Tools and Techniques for Software and System Health Management." Topics include: Ingredients of a Guidance, Navigation, and Control System (GN and C); Selected GN and C Testbed example; Health Management of major ingredients; ISWHM testbed architecture; and Conclusions and next Steps.

  8. Research on software systems dependability at the OECD Halden Reactor Project

    International Nuclear Information System (INIS)

    Sivertsen, Terje; Owre, Fridtjov

    2011-01-01

    Two central issues related to software systems dependability are those of safety integrity and safety demonstration. A proper understanding of these two issues are important for the selection of process, methods, techniques and tools to be used in the different life cycle phases of the software. Following a brief discussion on the concept of software safety integrity and its relationship to software systems dependability, this paper gives an introduction to research problems addressed by the OECD Halden Reactor Project within this area. The paper concludes with a discussion on the important role of safety demonstration in this context. (author)

  9. Application software for new BEPC interlock system

    International Nuclear Information System (INIS)

    Tang Shuming; Na Xiangyin; Chen Jiansong; Yu Yulan

    1997-01-01

    New BEPC (Beijing electron Positron collider) interlock system has been built in order to improve the reliability of personnel safety and interlock functions. Moreover, the system updates BEPC operation message once every 6 seconds, which are displayed on TV screens at the major entrances. Since March of 1996, new BEPC interlock system has been operating reliably. The hardware of the system is based on Programmable Logic Controllers (PLC). A multimedia IBM/PC-586 as the host computer of the PLCs, monitors the PLC system via serial port COM2. The PC communicates with the central computer VAX-4500 of BEPC control system and gets operating massage of the accelerator through serial port COM3. The application software on the host computer has been developed. Visual C++ for MS-Windows 3.2 TM is selected as the work bench. It provides nice tools for building programs, such as APP STUDIO, CLASS WIZARD, APP WIZARD and debugger tool. The author describes the design idea and the structure of the application software. Error tolerance is taken into consideration. The author also presents a small database and its data structure for the application

  10. Sources, occurrence and predicted aquatic impact of legacy and contemporary pesticides in streams

    International Nuclear Information System (INIS)

    McKnight, Ursula S.; Rasmussen, Jes J.; Kronvang, Brian; Binning, Philip J.; Bjerg, Poul L.

    2015-01-01

    We couple current findings of pesticides in surface and groundwater to the history of pesticide usage, focusing on the potential contribution of legacy pesticides to the predicted ecotoxicological impact on benthic macroinvertebrates in headwater streams. Results suggest that groundwater, in addition to precipitation and surface runoff, is an important source of pesticides (particularly legacy herbicides) entering surface water. In addition to current-use active ingredients, legacy pesticides, metabolites and impurities are important for explaining the estimated total toxicity attributable to pesticides. Sediment-bound insecticides were identified as the primary source for predicted ecotoxicity. Our results support recent studies indicating that highly sorbing chemicals contribute and even drive impacts on aquatic ecosystems. They further indicate that groundwater contaminated by legacy and contemporary pesticides may impact adjoining streams. Stream observations of soluble and sediment-bound pesticides are valuable for understanding the long-term fate of pesticides in aquifers, and should be included in stream monitoring programs. - Highlights: • Findings comprised a range of contemporary and banned legacy pesticides in streams. • Groundwater is a significant pathway for some herbicides entering streams. • Legacy pesticides increased predicted aquatic toxicity by four orders of magnitude. • Sediment-bound insecticides were identified as the primary source for ecotoxicity. • Stream monitoring programs should include legacy pesticides to assess impacts. - Legacy pesticides, particularly sediment-bound insecticides were identified as the primary source for predicted ecotoxicity impacting benthic macroinvertebrates in headwater streams

  11. Software configuration management plan, 241-AY and 241-AZ tank farm MICON automation system

    International Nuclear Information System (INIS)

    Hill, L.F.

    1997-01-01

    This document establishes a Computer Software Configuration Management Plan (CSCM) for controlling software for the MICON Distributed Control System (DCS) located at the 241-AY and 241-AZ Aging Waste Tank Farm facilities in the 200 East Area. The MICON DCS software controls and monitors the instrumentation and equipment associated with plant systems and processes. A CSCM identifies and defines the configuration items in a system (section 3.1), controls the release and change of these items throughout the system life cycle (section 3.2), records and reports the status of configuration items and change requests (section 3.3), and verifies the completeness and correctness of the items (section 3.4). All software development before initial release, or before software is baselined, is considered developmental. This plan does not apply to developmental software. This plan applies to software that has been baselined and released. The MICON software will monitor and control the related instrumentation and equipment of the 241-AY and 241-AZ Tank Farm ventilation systems. Eventually, this software may also assume the monitoring and control of the tank sludge washing equipment and other systems as they are brought on line. This plan applies to the System Cognizant Manager and MICON Cognizant Engineer (who is also referred to herein as the system administrator) responsible for the software/hardware and administration of the MICON system. This document also applies to any other organizations within Tank Farms which are currently active on the system including system cognizant engineers, nuclear operators, technicians, and control room supervisors

  12. U.S. Spacesuit Legacy: Maintaining it for the Future

    Science.gov (United States)

    Chullen, Cinda; McMann, Joe; Thomas, Ken; Kosmo, Joe; Lewis, Cathleen; Wright, Rebecca; Bitterly, Rose; Olivia, Vladenka Rose

    2013-01-01

    The history of U.S. spacesuit development and its use are rich with information on lessons learned, and constitutes a valuable legacy to those designing spacesuits for the future, as well as to educators, students, and the general public. The genesis of lessons learned is best understood by studying the evolution of past spacesuit programs - how the challenges and pressures of the times influenced the direction of the various spacesuit programs. This paper shows how the legacy of various spacesuit-related programs evolved in response to these forces. Important aspects of how this U.S. spacesuit legacy is being preserved today is described, including the archiving of spacesuit hardware, important documents, videos, oral history, and the rapidly expanding U.S. Spacesuit Knowledge Capture program.

  13. Discrete Address Beacon System (DABS) Software System Reliability Modeling and Prediction.

    Science.gov (United States)

    1981-06-01

    Service ( ATARS ) module because of its interim status. Reliability prediction models for software modules were derived and then verified by matching...System (A’iCR3BS) and thus can be introduced gradually and economically without ma jor olper- ational or procedural change. Since DABS uses monopulse...lineanaly- sis tools or are ured during maintenance or pre-initialization were not modeled because they are not part of the mission software. The ATARS

  14. Cassini: The Journey and the Legacy

    KAUST Repository

    Porco, Carolyn

    2018-01-15

    An international mission to explore, in depth, the Saturnian system ヨthe planet Saturn and its magnetosphere, glorious rings, and many moons- begun over 27 years ago. After seven years of development, the Cassini spacecraft was launched in 1997, spent seven years trekking to Saturn, and finally entered Saturn orbit in the summer of 2004. In the course of its 13 years orbiting this ring world, Cassini returned over 450 thousand images, 635GB of data, and invaluable insights on the solar systemメs most splendid and scientifically rich planetary system. In this lecture, Carolyn Porco, the leader of the imaging science team on NASA\\'s Cassini mission, will delight her audience with a retrospective look at what has been learned from this profoundly successful mission and what its final legacy is likely to be.

  15. Review of Bruce A reactor regulating system software

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1995-12-01

    Each of the four reactor units at the Ontario Hydro Bruce A Nuclear Generating Station is controlled by the Reactor Regulating System (RRS) software running on digital computers. This research report presents an assessment of the quality and reliability of the RRS software based on a review of the RRS design documentation, an analysis of certain significant Event Reports (SERs), and an examination of selected software changes. We found that the RRS software requirements (i.e., what the software should do) were never clearly documented, and that design documents, which should describe how the requirements are implemented, are incomplete and inaccurate. Some RRS-related SERs (i.e., reports on unexpected incidents relating to the reactor control) implied that there were faults in the RRS, or that RRS changes should be made to help prevent certain unexpected events. The follow-up investigations were generally poorly documented, and so it could not usually be determined that problems were properly resolved. The Bruce A software change control procedures require improvement. For the software changes examined, there was insufficient evidence provided by Ontario Hydro that the required procedures regarding change approval, independent review, documentation updates, and testing were followed. Ontario Hydro relies on the expertise of their technical staff to modify the RRS software correctly; they have confidence in the software code itself, even if the documentation is not up-to-date. Ontario Hydro did not produce the documentation required for an independent formal assessment of the reliability of the RRS. (author). 37 refs., 3 figs.

  16. Review of Bruce A reactor regulating system software

    International Nuclear Information System (INIS)

    1995-12-01

    Each of the four reactor units at the Ontario Hydro Bruce A Nuclear Generating Station is controlled by the Reactor Regulating System (RRS) software running on digital computers. This research report presents an assessment of the quality and reliability of the RRS software based on a review of the RRS design documentation, an analysis of certain significant Event Reports (SERs), and an examination of selected software changes. We found that the RRS software requirements (i.e., what the software should do) were never clearly documented, and that design documents, which should describe how the requirements are implemented, are incomplete and inaccurate. Some RRS-related SERs (i.e., reports on unexpected incidents relating to the reactor control) implied that there were faults in the RRS, or that RRS changes should be made to help prevent certain unexpected events. The follow-up investigations were generally poorly documented, and so it could not usually be determined that problems were properly resolved. The Bruce A software change control procedures require improvement. For the software changes examined, there was insufficient evidence provided by Ontario Hydro that the required procedures regarding change approval, independent review, documentation updates, and testing were followed. Ontario Hydro relies on the expertise of their technical staff to modify the RRS software correctly; they have confidence in the software code itself, even if the documentation is not up-to-date. Ontario Hydro did not produce the documentation required for an independent formal assessment of the reliability of the RRS. (author). 37 refs., 3 figs

  17. The use of prime implicants in dependability analysis of software controlled systems

    International Nuclear Information System (INIS)

    Yau, Michael; Apostolakis, George; Guarro, Sergio

    1998-01-01

    The behavior of software controlled systems is usually non-binary and dynamic. It is, thus, convenient to employ multi-valued logic to model these systems. Multi-valued logic functions can be used to represent the functional and temporal relationships between the software and hardware components. The resulting multi-valued logic model can be analyzed deductively, i.e. by tracking causality in reverse from undesirable 'top' events to identify faults that may be present in the system. The result of this deductive analysis is a set of prime implicants for a user-defined system top event. The prime implicants represent all the combinations of basic component conditions and software input conditions that may result in the top event; they are the extension to multi-valued logic of the concept of minimal cut sets that is used routinely in the analysis of binary fault trees. This paper discusses why prime implicants are needed in the dependability analysis of software controlled systems, how they are generated, and how they are used to identify faults in a software controlled system

  18. The use of prime implicants in dependability analysis of software controlled systems

    Energy Technology Data Exchange (ETDEWEB)

    Yau, Michael; Apostolakis, George; Guarro, Sergio

    1998-11-01

    The behavior of software controlled systems is usually non-binary and dynamic. It is, thus, convenient to employ multi-valued logic to model these systems. Multi-valued logic functions can be used to represent the functional and temporal relationships between the software and hardware components. The resulting multi-valued logic model can be analyzed deductively, i.e. by tracking causality in reverse from undesirable 'top' events to identify faults that may be present in the system. The result of this deductive analysis is a set of prime implicants for a user-defined system top event. The prime implicants represent all the combinations of basic component conditions and software input conditions that may result in the top event; they are the extension to multi-valued logic of the concept of minimal cut sets that is used routinely in the analysis of binary fault trees. This paper discusses why prime implicants are needed in the dependability analysis of software controlled systems, how they are generated, and how they are used to identify faults in a software controlled system.

  19. Past as Prologue: Educational Psychology's Legacy and Progeny

    Science.gov (United States)

    Alexander, Patricia A.

    2018-01-01

    On the occasion of the 125th anniversary of the American Psychological Association, the legacies and progenies of the discipline of educational psychology are explored. To capture those legacies, transformational and influential contributions by educational psychologists to schools and society are described as key themes. Those themes entail: the…

  20. The 5L Instructional Design For Exploring Legacies through Biography

    Science.gov (United States)

    Boulware, Beverly J.; Monroe, Eula E.; Wilcox, Bradley Ray

    2013-01-01

    People who have impacted generations have left legacies we can explore today through biographies. The 5L instructional design introduced in this article includes five components: Listen, Learn, Locate, Link, and Legacy. In the "Listen" section, teachers use storytelling and read-alouds to introduce individuals who shaped history. During…

  1. CRISP90 - SOFTWARE DESIGN ANALYZER SYSTEM

    Science.gov (United States)

    Tausworthe, R. C.

    1994-01-01

    The CRISP90 Software Design Analyzer System, an update of CRISP-80, is a set of programs forming a software design and documentation tool which supports top-down, hierarchic, modular, structured design and programming methodologies. The quality of a computer program can often be significantly influenced by the design medium in which the program is developed. The medium must foster the expression of the programmer's ideas easily and quickly, and it must permit flexible and facile alterations, additions, and deletions to these ideas as the design evolves. The CRISP90 software design analyzer system was developed to provide the PDL (Programmer Design Language) programmer with such a design medium. A program design using CRISP90 consists of short, English-like textual descriptions of data, interfaces, and procedures that are imbedded in a simple, structured, modular syntax. The display is formatted into two-dimensional, flowchart-like segments for a graphic presentation of the design. Together with a good interactive full-screen editor or word processor, the CRISP90 design analyzer becomes a powerful tool for the programmer. In addition to being a text formatter, the CRISP90 system prepares material that would be tedious and error prone to extract manually, such as a table of contents, module directory, structure (tier) chart, cross-references, and a statistics report on the characteristics of the design. Referenced modules are marked by schematic logic symbols to show conditional, iterative, and/or concurrent invocation in the program. A keyword usage profile can be generated automatically and glossary definitions inserted into the output documentation. Another feature is the capability to detect changes that were made between versions. Thus, "change-bars" can be placed in the output document along with a list of changed pages and a version history report. Also, items may be marked as "to be determined" and each will appear on a special table until the item is

  2. Scientific legacy of Stanley Ruby

    International Nuclear Information System (INIS)

    Shenoy, G. K.

    2006-01-01

    Stanley L. Ruby (1924-2004) made major contributions to Moessbauer spectroscopy and was the first to suggest the feasibility of observing the Moessbauer effect using synchrotron radiation. In this article we recall his scientific legacy that have inspired his scientific colleagues.

  3. Bridging the gap: linking a legacy hospital information system with a filmless radiology picture archiving and communications system within a nonhomogeneous environment.

    Science.gov (United States)

    Rubin, R K; Henri, C J; Cox, R D

    1999-05-01

    A health level 7 (HL7)-conformant data link to exchange information between the mainframe hospital information system (HIS) of our hospital and our home-grown picture archiving and communications system (PACS) is a result of a collaborative effort between the HIS department and the PACS development team. Based of the ability to link examination requisitions and image studies, applications have been generated to optimise workflow and to improve the reliability and distribution of radiology information. Now, images can be routed to individual radiologists and clinicians; worklists facilitate radiology reporting; applications exist to create, edit, and view reports and images via the internet; and automated quality control now limits the incidence of "lost" cases and errors in image routing. By following the HL7 standard to develop the gateway to the legacy system, the development of a radiology information system for booking, reading, reporting, and billing remains universal and does not preclude the option to integrate off-the-shelf commercial products.

  4. Subsystem software for TSTA [Tritium Systems Test Assembly

    International Nuclear Information System (INIS)

    Mann, L.W.; Claborn, G.W.; Nielson, C.W.

    1987-01-01

    The Subsystem Control Software at the Tritium System Test Assembly (TSTA) must control sophisticated chemical processes through the physical operation of valves, motor controllers, gas sampling devices, thermocouples, pressure transducers, and similar devices. Such control software has to be capable of passing stringent quality assurance (QA) criteria to provide for the safe handling of significant amounts of tritium on a routine basis. Since many of the chemical processes and physical components are experimental, the control software has to be flexible enough to allow for trial/error learning curve, but still protect the environment and personnel from exposure to unsafe levels of radiation. The software at TSTA is implemented in several levels as described in a preceding paper in these proceedings. This paper depends on information given in the preceding paper for understanding. The top level is the Subsystem Control level

  5. KAERI software verification and validation guideline for developing safety-critical software in digital I and C system of NPP

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jang Yeol; Lee, Jang Soo; Eom, Heung Seop

    1997-07-01

    This technical report is to present V and V guideline development methodology for safety-critical software in NPP safety system. Therefore it is to present V and V guideline of planning phase for the NPP safety system in addition to critical safety items, for example, independence philosophy, software safety analysis concept, commercial off the shelf (COTS) software evaluation criteria, inter-relationships between other safety assurance organizations, including the concepts of existing industrial standard, IEEE Std-1012, IEEE Std-1059. This technical report includes scope of V and V guideline, guideline framework as part of acceptance criteria, V and V activities and task entrance as part of V and V activity and exit criteria, review and audit, testing and QA records of V and V material and configuration management, software verification and validation plan production etc., and safety-critical software V and V methodology. (author). 11 refs.

  6. KAERI software verification and validation guideline for developing safety-critical software in digital I and C system of NPP

    International Nuclear Information System (INIS)

    Kim, Jang Yeol; Lee, Jang Soo; Eom, Heung Seop.

    1997-07-01

    This technical report is to present V and V guideline development methodology for safety-critical software in NPP safety system. Therefore it is to present V and V guideline of planning phase for the NPP safety system in addition to critical safety items, for example, independence philosophy, software safety analysis concept, commercial off the shelf (COTS) software evaluation criteria, inter-relationships between other safety assurance organizations, including the concepts of existing industrial standard, IEEE Std-1012, IEEE Std-1059. This technical report includes scope of V and V guideline, guideline framework as part of acceptance criteria, V and V activities and task entrance as part of V and V activity and exit criteria, review and audit, testing and QA records of V and V material and configuration management, software verification and validation plan production etc., and safety-critical software V and V methodology. (author). 11 refs

  7. Social software: E-learning beyond learning management systems

    DEFF Research Database (Denmark)

    Dalsgaard, Christian

    2006-01-01

    The article argues that it is necessary to move e-learning beyond learning management systems and engage students in an active use of the web as a resource for their self-governed, problem-based and collaborative activities. The purpose of the article is to discuss the potential of social software...... to move e-learning beyond learning management systems. An approach to use of social software in support of a social constructivist approach to e-learning is presented, and it is argued that learning management systems do not support a social constructivist approach which emphasizes self-governed learning...... activities of students. The article suggests a limitation of the use of learning management systems to cover only administrative issues. Further, it is argued that students' self-governed learning processes are supported by providing students with personal tools and engaging them in different kinds of social...

  8. Software and system development using virtual platforms full-system simulation with wind river simics

    CERN Document Server

    Aarno, Daniel

    2014-01-01

    Virtual platforms are finding widespread use in both pre- and post-silicon computer software and system development. They reduce time to market, improve system quality, make development more efficient, and enable truly concurrent hardware/software design and bring-up. Virtual platforms increase productivity with unparalleled inspection, configuration, and injection capabilities. In combination with other types of simulators, they provide full-system simulations where computer systems can be tested together with the environment in which they operate. This book is not only about what simulat

  9. An autonomous sensor module based on a legacy CCTV camera

    Science.gov (United States)

    Kent, P. J.; Faulkner, D. A. A.; Marshall, G. F.

    2016-10-01

    A UK MoD funded programme into autonomous sensors arrays (SAPIENT) has been developing new, highly capable sensor modules together with a scalable modular architecture for control and communication. As part of this system there is a desire to also utilise existing legacy sensors. The paper reports upon the development of a SAPIENT-compliant sensor module using a legacy Close-Circuit Television (CCTV) pan-tilt-zoom (PTZ) camera. The PTZ camera sensor provides three modes of operation. In the first mode, the camera is automatically slewed to acquire imagery of a specified scene area, e.g. to provide "eyes-on" confirmation for a human operator or for forensic purposes. In the second mode, the camera is directed to monitor an area of interest, with zoom level automatically optimized for human detection at the appropriate range. Open source algorithms (using OpenCV) are used to automatically detect pedestrians; their real world positions are estimated and communicated back to the SAPIENT central fusion system. In the third mode of operation a "follow" mode is implemented where the camera maintains the detected person within the camera field-of-view without requiring an end-user to directly control the camera with a joystick.

  10. A software system for oilfield facility investment minimization

    International Nuclear Information System (INIS)

    Ding, Z.X.; Startzman, R.A.

    1996-01-01

    Minimizing investment in oilfield development is an important subject that has attracted a considerable amount of industry attention. One method to reduce investment involves the optimal placement and selection of production facilities. Because of the large amount of capital used in this process, saving a small percent of the total investment may represent a large monetary value. The literature reports algorithms using mathematical programming techniques that were designed to solve the proposed problem in a global optimal manner. Owing to the high-computational complexity and the lack of user-friendly interfaces for data entry and results display, mathematical programming techniques have not been given enough attention in practice. This paper describes an interactive, graphical software system that provides a global optimal solution to the problem of placement and selection of production facilities in oil-field development processes. This software system can be used as an investment minimization tool and a scenario-study simulator. The developed software system consists of five basic modules: (1) an interactive data-input unit, (2) a cost function generator, (3) an optimization unit, (4) a graphic-output display, and (5) a sensitivity-analysis unit

  11. Managing a project's legacy: implications for organizations and project management

    Science.gov (United States)

    Cooper, Lynne P.; Hecht, Michael H.; Majchrzak, Ann

    2003-01-01

    Organizations that rely on projects to implement their products must find effective mechanisms for propagating lessons learned on one project throughout the organization. A broad view of what constitutes a project's 'legacy' is presented that includes not just the design products and leftover parts, but new processes, relationships, technology, skills, planning data, and performance metrics. Based on research evaluating knowledge reuse in innovative contexts, this paper presents an approach to project legacy management that focuses on collecting and using legacy knowledge to promote organizational learning and effective reuse, while addressing factors of post-project responsibility, information obsolescence, and the importance of ancillary contextual information. .

  12. Main Principles of the Organization of Decommissioning Activities for Legacy Sites

    Energy Technology Data Exchange (ETDEWEB)

    Mikheykin, S.V., E-mail: Mikheykin@rosrao.ru [Department of Techniques for Remediation, Federal State Unitary Enterprise ' RosRAO' , Moscow (Russian Federation)

    2013-08-15

    As a result of more than 60 years development of nuclear industry in the former Soviet Union and in the Russian Federation there has accumulated a number of unresolved problems associated with contamination of facilities and environment during the early stages of research and industrial activities. Prior to the year 2000 most of the problems were solved slowly; the main decisions were postponed for the future. During that time were done the local works for the rehabilitation of contaminated sites. The Federal Target Programme ''Nuclear and Radiation Safety for 2008 and for the period to 2015'' was adopted in 2008. Analysis of accumulated experience as result of previous work on decontamination to develop new project management system for the rehabilitation of the nuclear legacy is needed. This CRP contribution is aimed at solving the tasks of the rehabilitation of the nuclear legacy. (author)

  13. Towards automated construction of dependable software/hardware systems

    Energy Technology Data Exchange (ETDEWEB)

    Yakhnis, A.; Yakhnis, V. [Pioneer Technologies & Rockwell Science Center, Albuquerque, NM (United States)

    1997-11-01

    This report contains viewgraphs on the automated construction of dependable computer architecture systems. The outline of this report is: examples of software/hardware systems; dependable systems; partial delivery of dependability; proposed approach; removing obstacles; advantages of the approach; criteria for success; current progress of the approach; and references.

  14. Reliable Software Development for Machine Protection Systems

    CERN Document Server

    Anderson, D; Dragu, M; Fuchsberger, K; Garnier, JC; Gorzawski, AA; Koza, M; Krol, K; Misiowiec, K; Stamos, K; Zerlauth, M

    2014-01-01

    The Controls software for the Large Hadron Collider (LHC) at CERN, with more than 150 millions lines of code, resides amongst the largest known code bases in the world1. Industry has been applying Agile software engineering techniques for more than two decades now, and the advantages of these techniques can no longer be ignored to manage the code base for large projects within the accelerator community. Furthermore, CERN is a particular environment due to the high personnel turnover and manpower limitations, where applying Agile processes can improve both, the codebase management as well as its quality. This paper presents the successful application of the Agile software development process Scrum for machine protection systems at CERN, the quality standards and infrastructure introduced together with the Agile process as well as the challenges encountered to adapt it to the CERN environment.

  15. Surveillance Analysis Computer System (SACS): Software requirements specification (SRS). Revision 2

    International Nuclear Information System (INIS)

    Glasscock, J.A.

    1995-01-01

    This document is the primary document establishing requirements for the Surveillance Analysis Computer System (SACS) database, an Impact Level 3Q system. SACS stores information on tank temperatures, surface levels, and interstitial liquid levels. This information is retrieved by the customer through a PC-based interface and is then available to a number of other software tools. The software requirements specification (SRS) describes the system requirements for the SACS Project, and follows the Standard Engineering Practices (WHC-CM-6-1), Software Practices (WHC-CM-3-10) and Quality Assurance (WHC-CM-4-2, QR 19.0) policies

  16. Software for microcircuit systems

    International Nuclear Information System (INIS)

    Kunz, P.F.

    1978-10-01

    Modern Large Scale Integration (LSI) microcircuits are meant to be programed in order to control the function that they perform. The basics of microprograming and new microcircuits have already been discussed. In this course, the methods of developing software for these microcircuits are explored. This generally requires a package of support software in order to assemble the microprogram, and also some amount of support software to test the microprograms and to test the microprogramed circuit itself. 15 figures, 2 tables

  17. Software development methodology for computer based I&C systems of prototype fast breeder reactor

    International Nuclear Information System (INIS)

    Manimaran, M.; Shanmugam, A.; Parimalam, P.; Murali, N.; Satya Murty, S.A.V.

    2015-01-01

    Highlights: • Software development methodology adopted for computer based I&C systems of PFBR is detailed. • Constraints imposed as part of software requirements and coding phase are elaborated. • Compliance to safety and security requirements are described. • Usage of CASE (Computer Aided Software Engineering) tools during software design, analysis and testing phase are explained. - Abstract: Prototype Fast Breeder Reactor (PFBR) is sodium cooled reactor which is in the advanced stage of construction in Kalpakkam, India. Versa Module Europa bus based Real Time Computer (RTC) systems are deployed for Instrumentation & Control of PFBR. RTC systems have to perform safety functions within the stipulated time which calls for highly dependable software. Hence, well defined software development methodology is adopted for RTC systems starting from the requirement capture phase till the final validation of the software product. V-model is used for software development. IEC 60880 standard and AERB SG D-25 guideline are followed at each phase of software development. Requirements documents and design documents are prepared as per IEEE standards. Defensive programming strategies are followed for software development using C language. Verification and validation (V&V) of documents and software are carried out at each phase by independent V&V committee. Computer aided software engineering tools are used for software modelling, checking for MISRA C compliance and to carry out static and dynamic analysis. Various software metrics such as cyclomatic complexity, nesting depth and comment to code are checked. Test cases are generated using equivalence class partitioning, boundary value analysis and cause and effect graphing techniques. System integration testing is carried out wherein functional and performance requirements of the system are monitored

  18. Software development methodology for computer based I&C systems of prototype fast breeder reactor

    Energy Technology Data Exchange (ETDEWEB)

    Manimaran, M., E-mail: maran@igcar.gov.in; Shanmugam, A.; Parimalam, P.; Murali, N.; Satya Murty, S.A.V.

    2015-10-15

    Highlights: • Software development methodology adopted for computer based I&C systems of PFBR is detailed. • Constraints imposed as part of software requirements and coding phase are elaborated. • Compliance to safety and security requirements are described. • Usage of CASE (Computer Aided Software Engineering) tools during software design, analysis and testing phase are explained. - Abstract: Prototype Fast Breeder Reactor (PFBR) is sodium cooled reactor which is in the advanced stage of construction in Kalpakkam, India. Versa Module Europa bus based Real Time Computer (RTC) systems are deployed for Instrumentation & Control of PFBR. RTC systems have to perform safety functions within the stipulated time which calls for highly dependable software. Hence, well defined software development methodology is adopted for RTC systems starting from the requirement capture phase till the final validation of the software product. V-model is used for software development. IEC 60880 standard and AERB SG D-25 guideline are followed at each phase of software development. Requirements documents and design documents are prepared as per IEEE standards. Defensive programming strategies are followed for software development using C language. Verification and validation (V&V) of documents and software are carried out at each phase by independent V&V committee. Computer aided software engineering tools are used for software modelling, checking for MISRA C compliance and to carry out static and dynamic analysis. Various software metrics such as cyclomatic complexity, nesting depth and comment to code are checked. Test cases are generated using equivalence class partitioning, boundary value analysis and cause and effect graphing techniques. System integration testing is carried out wherein functional and performance requirements of the system are monitored.

  19. Network, system, and status software enhancements for the autonomously managed electrical power system breadboard. Volume 1: Project summary

    Science.gov (United States)

    Mckee, James W.

    1990-01-01

    This volume (1 of 4) gives a summary of the original AMPS software system configuration, points out some of the problem areas in the original software design that this project is to address, and in the appendix collects all the bimonthly status reports. The purpose of AMPS is to provide a self reliant system to control the generation and distribution of power in the space station. The software in the AMPS breadboard can be divided into three levels: the operating environment software, the protocol software, and the station specific software. This project deals only with the operating environment software and the protocol software. The present station specific software will not change except as necessary to conform to new data formats.

  20. (Resurveying Mediterranean Rural Landscapes: GIS and Legacy Survey Data

    Directory of Open Access Journals (Sweden)

    Robert Witcher

    2008-07-01

    Full Text Available Legacy data have always been important for Mediterranean archaeologists. Over the past decade, one specific category of legacy data, that deriving from regional survey, has become particularly important. Not only has the scale of research questions become larger (requiring greater reliance on others' data, but the surface archaeological record is deteriorating (diminishing the ability to recover good data. The legacy data from many individual surveys have now been subject to digitisation and GIS analysis, successfully redeploying data collected for one purpose within new theoretical and interpretive frameworks. However, a key research focus is now comparative survey - using the results of many different Mediterranean surveys side-by-side to identify regional variability in settlement organisation, economy and demography. In order to overcome the significant methodological differences between these surveys, attention has focused on the documentation of metadata. Yet, many legacy data lack vital information about their creation and hence how they might be (reinterpreted and compared. GIS has been advanced as an environment in which to contain, order and analyse the data necessary for comparative survey. However, there is a danger that the technology will facilitate inappropriate use of these datasets in a way that fails to acknowledge and understand the very real differences between them. Here, emphasis is placed upon the use of GIS as a space for exploratory data analysis: a process that encompasses and emphasises the integral processes of digitisation, visualisation and simple analysis for the characterisation of datasets in order to derive an alternative form of metadata. Particular emphasis is placed upon the interaction of past human behaviour (e.g. in the Roman period and archaeological recovery (i.e. the behaviour of archaeologists in the present, or recent past; these two sets of 'social action' combine to create distinctive archaeological

  1. Software Development for the Hobby-Eberly Telescope's Segment Alignment Maintenance System using LABView

    Science.gov (United States)

    Hall, Drew P.; Ly, William; Howard, Richard T.; Weir, John; Rakoczy, John; Roe, Fred (Technical Monitor)

    2002-01-01

    The software development for an upgrade to the Hobby-Eberly Telescope (HET) was done in LABView. In order to improve the performance of the HET at the McDonald Observatory, a closed-loop system had to be implemented to keep the mirror segments aligned during periods of observation. The control system, called the Segment Alignment Maintenance System (SAMs), utilized inductive sensors to measure the relative motions of the mirror segments. Software was developed in LABView to tie the sensors, operator interface, and mirror-control motors together. Developing the software in LABView allowed the system to be flexible, understandable, and able to be modified by the end users. Since LABView is built using block diagrams, the software naturally followed the designed control system's block and flow diagrams, and individual software blocks could be easily verified. LABView's many built-in display routines allowed easy visualization of diagnostic and health-monitoring data during testing. Also, since LABView is a multi-platform software package, different programmers could develop the code remotely on various types of machines. LABView s ease of use facilitated rapid prototyping and field testing. There were some unanticipated difficulties in the software development, but the use of LABView as the software "language" for the development of SAMs contributed to the overall success of the project.

  2. 2014 Runtime Systems Summit. Runtime Systems Report

    Energy Technology Data Exchange (ETDEWEB)

    Sarkar, Vivek [US Dept. of Energy, Washington, DC (United States); Budimlic, Zoran [US Dept. of Energy, Washington, DC (United States); Kulkani, Milind [US Dept. of Energy, Washington, DC (United States)

    2016-09-19

    This report summarizes runtime system challenges for exascale computing, that follow from the fundamental challenges for exascale systems that have been well studied in past reports, e.g., [6, 33, 34, 32, 24]. Some of the key exascale challenges that pertain to runtime systems include parallelism, energy efficiency, memory hierarchies, data movement, heterogeneous processors and memories, resilience, performance variability, dynamic resource allocation, performance portability, and interoperability with legacy code. In addition to summarizing these challenges, the report also outlines different approaches to addressing these significant challenges that have been pursued by research projects in the DOE-sponsored X-Stack and OS/R programs. Since there is often confusion as to what exactly the term “runtime system” refers to in the software stack, we include a section on taxonomy to clarify the terminology used by participants in these research projects. In addition, we include a section on deployment opportunities for vendors and government labs to build on the research results from these projects. Finally, this report is also intended to provide a framework for discussing future research and development investments for exascale runtime systems, and for clarifying the role of runtime systems in exascale software.

  3. A software Event Summation System for MDSplus

    International Nuclear Information System (INIS)

    Davis, W.M.; Mastrovito, D.M.; Roney, P.G.; Sichta, P.

    2008-01-01

    The MDSplus data acquisition and management system uses software events for communication among interdependent processes anywhere on the network. Actions can then be triggered, such as a data-acquisition routine, or analysis or display programs waiting for data. A small amount of data, such as a shot number, can be passed with these events. Since programs sometimes need more than one data set, we developed a system on NSTX to declare composite events using logical AND and OR operations. The system is written in the IDL language, so it can be run on Linux, Macintosh or Windows platforms. Like MDSplus, the Experimental Physics and Industrial Control System (EPICS) is a core component of the NSTX software environment. The Event Summation System provides an IDL-based interface to EPICS. This permits EPICS-aware processes to be synchronized with MDSplus-aware processes, to provide, for example, engineering operators information about physics data acquisition and analysis. Reliability was a more important design consideration than performance for this system; the system's architecture includes features to support this. The system has run for weeks at a time without requiring manual intervention. Hundreds of incoming events per second can be handled reliably. All incoming and declared events are logged with a timestamp. The system can be configured easily through a single, easy-to-read text file

  4. Computer software.

    Science.gov (United States)

    Rosenthal, L E

    1986-10-01

    Software is the component in a computer system that permits the hardware to perform the various functions that a computer system is capable of doing. The history of software and its development can be traced to the early nineteenth century. All computer systems are designed to utilize the "stored program concept" as first developed by Charles Babbage in the 1850s. The concept was lost until the mid-1940s, when modern computers made their appearance. Today, because of the complex and myriad tasks that a computer system can perform, there has been a differentiation of types of software. There is software designed to perform specific business applications. There is software that controls the overall operation of a computer system. And there is software that is designed to carry out specialized tasks. Regardless of types, software is the most critical component of any computer system. Without it, all one has is a collection of circuits, transistors, and silicone chips.

  5. Software architecture for the ORNL large coil test facility data system

    International Nuclear Information System (INIS)

    Blair, E.T.; Baylor, L.R.

    1986-01-01

    The VAX based data acquisition system for the international fusion superconducting magnetic test facility (IFSMTF) at Oak Ridge National Laboratory (ORNL) is a second generation system that evolved from a PDP-11/60 based system used during the initial phase of facility testing. The VAX based software represents a layered implementation that provides integrated access to all of the data sources within the system, decoupling en-user data retrieval from various front-end data sources through a combination of software architecture and instrumentation data bases. Independent VAX processes manage the various front-end data sources, each being responsible for controlling, monitoring, acquiring, and disposing data and control parameters for access from the data retrieval software

  6. Nitrate retention capacity of milldam-impacted legacy sediments and relict A horizon soils

    Directory of Open Access Journals (Sweden)

    J. N. Weitzman

    2017-05-01

    Full Text Available While eutrophication is often attributed to contemporary nutrient pollution, there is growing evidence that past practices, like the accumulation of legacy sediment behind historic milldams, are also important. Given their prevalence, there is a critical need to understand how N flows through, and is retained in, legacy sediments to improve predictions and management of N transport from uplands to streams in the context of climatic variability and land-use change. Our goal was to determine how nitrate (NO3− is cycled through the soil of a legacy-sediment-strewn stream before and after soil drying. We extracted 10.16 cm radius intact soil columns that extended 30 cm into each of the three significant soil horizons at Big Spring Run (BSR in Lancaster, Pennsylvania: surface legacy sediment characterized by a newly developing mineral A horizon soil, mid-layer legacy sediment consisting of mineral B horizon soil and a dark, organic-rich, buried relict A horizon soil. Columns were first preincubated at field capacity and then isotopically labeled nitrate (15NO3− was added and allowed to drain to estimate retention. The columns were then air-dried and subsequently rewet with N-free water and allowed to drain to quantify the drought-induced loss of 15NO3− from the different horizons. We found the highest initial 15N retention in the mid-layer legacy sediment (17 ± 4 % and buried relict A soil (14 ± 3 % horizons, with significantly lower retention in the surface legacy sediment (6 ± 1 % horizon. As expected, rewetting dry soil resulted in 15N losses in all horizons, with the greatest losses in the buried relict A horizon soil, followed by the mid-layer legacy sediment and surface legacy sediment horizons. The 15N remaining in the soil following the post-drought leaching was highest in the mid-layer legacy sediment, intermediate in the surface legacy sediment, and lowest in the buried relict A horizon soil. Fluctuations

  7. Nitrate retention capacity of milldam-impacted legacy sediments and relict A horizon soils

    Science.gov (United States)

    Weitzman, Julie N.; Kaye, Jason P.

    2017-05-01

    While eutrophication is often attributed to contemporary nutrient pollution, there is growing evidence that past practices, like the accumulation of legacy sediment behind historic milldams, are also important. Given their prevalence, there is a critical need to understand how N flows through, and is retained in, legacy sediments to improve predictions and management of N transport from uplands to streams in the context of climatic variability and land-use change. Our goal was to determine how nitrate (NO3-) is cycled through the soil of a legacy-sediment-strewn stream before and after soil drying. We extracted 10.16 cm radius intact soil columns that extended 30 cm into each of the three significant soil horizons at Big Spring Run (BSR) in Lancaster, Pennsylvania: surface legacy sediment characterized by a newly developing mineral A horizon soil, mid-layer legacy sediment consisting of mineral B horizon soil and a dark, organic-rich, buried relict A horizon soil. Columns were first preincubated at field capacity and then isotopically labeled nitrate (15NO3-) was added and allowed to drain to estimate retention. The columns were then air-dried and subsequently rewet with N-free water and allowed to drain to quantify the drought-induced loss of 15NO3- from the different horizons. We found the highest initial 15N retention in the mid-layer legacy sediment (17 ± 4 %) and buried relict A soil (14 ± 3 %) horizons, with significantly lower retention in the surface legacy sediment (6 ± 1 %) horizon. As expected, rewetting dry soil resulted in 15N losses in all horizons, with the greatest losses in the buried relict A horizon soil, followed by the mid-layer legacy sediment and surface legacy sediment horizons. The 15N remaining in the soil following the post-drought leaching was highest in the mid-layer legacy sediment, intermediate in the surface legacy sediment, and lowest in the buried relict A horizon soil. Fluctuations in the water table at BSR which affect

  8. Development of requirements tracking and verification system for the software design of distributed control system

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Chul Hwan; Kim, Jang Yeol; Kim, Jung Tack; Lee, Jang Soo; Ham, Chang Shik [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1999-12-31

    In this paper a prototype of Requirement Tracking and Verification System(RTVS) for a Distributed Control System was implemented and tested. The RTVS is a software design and verification tool. The main functions required by the RTVS are managing, tracking and verification of the software requirements listed in the documentation of the DCS. The analysis of DCS software design procedures and interfaces with documents were performed to define the user of the RTVS, and the design requirements for RTVS were developed. 4 refs., 3 figs. (Author)

  9. Development of requirements tracking and verification system for the software design of distributed control system

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Chul Hwan; Kim, Jang Yeol; Kim, Jung Tack; Lee, Jang Soo; Ham, Chang Shik [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1998-12-31

    In this paper a prototype of Requirement Tracking and Verification System(RTVS) for a Distributed Control System was implemented and tested. The RTVS is a software design and verification tool. The main functions required by the RTVS are managing, tracking and verification of the software requirements listed in the documentation of the DCS. The analysis of DCS software design procedures and interfaces with documents were performed to define the user of the RTVS, and the design requirements for RTVS were developed. 4 refs., 3 figs. (Author)

  10. Effects of Management Legacies on Stream Fish and Aquatic Benthic Macroinvertebrate Assemblages

    Science.gov (United States)

    Quist, Michael C.; Schultz, Randall D.

    2014-09-01

    Fish and benthic macroinvertebrate assemblages often provide insight on ecological conditions for guiding management actions. Unfortunately, land use and management legacies can constrain the structure of biotic communities such that they fail to reflect habitat quality. The purpose of this study was to describe patterns in fish and benthic macroinvertebrate assemblage structure, and evaluate relationships between biota and habitat characteristics in the Chariton River system of south-central Iowa, a system likely influenced by various potential management legacies (e.g., dams, chemical removal of fishes). We sampled fishes, benthic macroinvertebrates, and physical habitat from a total of 38 stream reaches in the Chariton River watershed during 2002-2005. Fish and benthic macroinvertebrate assemblages were dominated by generalist species tolerant of poor habitat quality; assemblages failed to show any apparent patterns with regard to stream size or longitudinal location within the watershed. Metrics used to summarize fish assemblages and populations [e.g., presence-absence, relative abundance, Index of Biotic Integrity for fish (IBIF)] were not related to habitat characteristics, except that catch rates of piscivores were positively related to the depth and the amount of large wood. In contrast, family richness of benthic macroinvertebrates, richness of Ephemeroptera, Trichoptera, and Plecoptera taxa, and IBI values for benthic macroinvertebrates (IBIBM) were positively correlated with the amount of overhanging vegetation and inversely related to the percentage of fine substrate. A long history of habitat alteration by row-crop agriculture and management legacies associated with reservoir construction has likely resulted in a fish assemblage dominated by tolerant species. Intolerant and sensitive fish species have not recolonized streams due to downstream movement barriers (i.e., dams). In contrast, aquatic insect assemblages reflected aquatic habitat, particularly

  11. Radioisotope thermoelectric generator transportation system subsystem 143 software development plan

    International Nuclear Information System (INIS)

    King, D.A.

    1994-01-01

    This plan describes the activities to be performed and the controls to be applied to the process of specifying, developing, and qualifying the data acquisition software for the Radioisotope Thermoelectric Generator (RTG) Transportation System Subsystem 143 Instrumentation and Data Acquisition System (IDAS). This plan will serve as a software quality assurance plan, a verification and validation (V and V) plan, and a configuration management plan

  12. Using CASE Software to Teach Undergraduates Systems Analysis and Design.

    Science.gov (United States)

    Wilcox, Russell E.

    1988-01-01

    Describes the design and delivery of a college course for information system students utilizing a Computer-Aided Software Engineering program. Discusses class assignments, cooperative learning, student attitudes, and the advantages of using this software in the course. (CW)

  13. A New Control System Software for SANS BATAN Spectrometer in Serpong, Indonesia

    International Nuclear Information System (INIS)

    Bharoto; Putra, Edy Giri Rachman

    2010-01-01

    The original main control system of the 36 meter small-angle neutron scattering (SANS) BATAN Spectrometer (SMARTer) has been replaced with the new ones due to the malfunction of the main computer. For that reason, a new control system software for handling all the control systems was also developed in order to put the spectrometer back in operation. The developed software is able to control the system such as rotation movement of six pinholes system, vertical movement of four neutron guide system with the total length of 16.5 m, two-directional movement of a neutron beam stopper, forward-backward movement of a 2D position sensitive detector (2D-PSD) along 16.7 m, etc. A Visual Basic language program running on Windows operating system was employed to develop the software and it can be operated by other remote computers in the local area network. All device positions and command menu are displayed graphically in the main monitor or window and each device control can be executed by clicking the control button. Those advantages are necessary required for developing a new user-friendly control system software. Finally, the new software has been tested for handling a complete SANS experiment and it works properly.

  14. Software in windows for staple compounding system of microcomputer nuclear mass scale

    International Nuclear Information System (INIS)

    Wang Yanting; Zhang Yongming; Wang Yu; Jin Dongping

    1998-01-01

    The software exploited in windows for staple compounding system of microcomputer nuclear mass scale is described. The staple compounding system is briefly narrated. The software structure and its realizing method are given

  15. Initial Comparison of Direct and Legacy Modeling Approaches for Radial Core Expansion Analysis

    International Nuclear Information System (INIS)

    Shemon, Emily R.

    2016-01-01

    Radial core expansion in sodium-cooled fast reactors provides an important reactivity feedback effect. As the reactor power increases due to normal start up conditions or accident scenarios, the core and surrounding materials heat up, causing both grid plate expansion and bowing of the assembly ducts. When the core restraint system is designed correctly, the resulting structural deformations introduce negative reactivity which decreases the reactor power. Historically, an indirect procedure has been used to estimate the reactivity feedback due to structural deformation which relies upon perturbation theory and coupling legacy physics codes with limited geometry capabilities. With advancements in modeling and simulation, radial core expansion phenomena can now be modeled directly, providing an assessment of the accuracy of the reactivity feedback coefficients generated by indirect legacy methods. Recently a new capability was added to the PROTEUS-SN unstructured geometry neutron transport solver to analyze deformed meshes quickly and directly. By supplying the deformed mesh in addition to the base configuration input files, PROTEUS-SN automatically processes material adjustments including calculation of region densities to conserve mass, calculation of isotopic densities according to material models (for example, sodium density as a function of temperature), and subsequent re-homogenization of materials. To verify the new capability of directly simulating deformed meshes, PROTEUS-SN was used to compute reactivity feedback for a series of contrived yet representative deformed configurations for the Advanced Burner Test Reactor design. The indirect legacy procedure was also performed to generate reactivity feedback coefficients for the same deformed configurations. Interestingly, the legacy procedure consistently overestimated reactivity feedbacks by 35% compared to direct simulations by PROTEUS-SN. This overestimation indicates that the legacy procedures are in fact

  16. Initial Comparison of Direct and Legacy Modeling Approaches for Radial Core Expansion Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Shemon, Emily R. [Argonne National Lab. (ANL), Argonne, IL (United States)

    2016-10-10

    Radial core expansion in sodium-cooled fast reactors provides an important reactivity feedback effect. As the reactor power increases due to normal start up conditions or accident scenarios, the core and surrounding materials heat up, causing both grid plate expansion and bowing of the assembly ducts. When the core restraint system is designed correctly, the resulting structural deformations introduce negative reactivity which decreases the reactor power. Historically, an indirect procedure has been used to estimate the reactivity feedback due to structural deformation which relies upon perturbation theory and coupling legacy physics codes with limited geometry capabilities. With advancements in modeling and simulation, radial core expansion phenomena can now be modeled directly, providing an assessment of the accuracy of the reactivity feedback coefficients generated by indirect legacy methods. Recently a new capability was added to the PROTEUS-SN unstructured geometry neutron transport solver to analyze deformed meshes quickly and directly. By supplying the deformed mesh in addition to the base configuration input files, PROTEUS-SN automatically processes material adjustments including calculation of region densities to conserve mass, calculation of isotopic densities according to material models (for example, sodium density as a function of temperature), and subsequent re-homogenization of materials. To verify the new capability of directly simulating deformed meshes, PROTEUS-SN was used to compute reactivity feedback for a series of contrived yet representative deformed configurations for the Advanced Burner Test Reactor design. The indirect legacy procedure was also performed to generate reactivity feedback coefficients for the same deformed configurations. Interestingly, the legacy procedure consistently overestimated reactivity feedbacks by 35% compared to direct simulations by PROTEUS-SN. This overestimation indicates that the legacy procedures are in fact

  17. A Software Defined Radio Based Airplane Communication Navigation Simulation System

    Science.gov (United States)

    He, L.; Zhong, H. T.; Song, D.

    2018-01-01

    Radio communication and navigation system plays important role in ensuring the safety of civil airplane in flight. Function and performance should be tested before these systems are installed on-board. Conventionally, a set of transmitter and receiver are needed for each system, thus all the equipment occupy a lot of space and are high cost. In this paper, software defined radio technology is applied to design a common hardware communication and navigation ground simulation system, which can host multiple airplane systems with different operating frequency, such as HF, VHF, VOR, ILS, ADF, etc. We use a broadband analog frontend hardware platform, universal software radio peripheral (USRP), to transmit/receive signal of different frequency band. Software is compiled by LabVIEW on computer, which interfaces with USRP through Ethernet, and is responsible for communication and navigation signal processing and system control. An integrated testing system is established to perform functional test and performance verification of the simulation signal, which demonstrate the feasibility of our design. The system is a low-cost and common hardware platform for multiple airplane systems, which provide helpful reference for integrated avionics design.

  18. The SOFIA Mission Control System Software

    Science.gov (United States)

    Heiligman, G. M.; Brock, D. R.; Culp, S. D.; Decker, P. H.; Estrada, J. C.; Graybeal, J. B.; Nichols, D. M.; Paluzzi, P. R.; Sharer, P. J.; Pampell, R. J.; Papke, B. L.; Salovich, R. D.; Schlappe, S. B.; Spriestersbach, K. K.; Webb, G. L.

    1999-05-01

    The Stratospheric Observatory for Infrared Astronomy (SOFIA) will be delivered with a computerized mission control system (MCS). The MCS communicates with the aircraft's flight management system and coordinates the operations of the telescope assembly, mission-specific subsystems, and the science instruments. The software for the MCS must be reliable and flexible. It must be easily usable by many teams of observers with widely differing needs, and it must support non-intrusive access for education and public outreach. The technology must be appropriate for SOFIA's 20-year lifetime. The MCS software development process is an object-oriented, use case driven approach. The process is iterative: delivery will be phased over four "builds"; each build will be the result of many iterations; and each iteration will include analysis, design, implementation, and test activities. The team is geographically distributed, coordinating its work via Web pages, teleconferences, T.120 remote collaboration, and CVS (for Internet-enabled configuration management). The MCS software architectural design is derived in part from other observatories' experience. Some important features of the MCS are: * distributed computing over several UNIX and VxWorks computers * fast throughput of time-critical data * use of third-party components, such as the Adaptive Communications Environment (ACE) and the Common Object Request Broker Architecture (CORBA) * extensive configurability via stored, editable configuration files * use of several computer languages so developers have "the right tool for the job". C++, Java, scripting languages, Interactive Data Language (from Research Systems, Int'l.), XML, and HTML will all be used in the final deliverables. This paper reports on work in progress, with the final product scheduled for delivery in 2001. This work was performed for Universities Space Research Association for NASA under contract NAS2-97001.

  19. Bereaved mothers' and fathers' perceptions of a legacy intervention for parents of infants in the NICU.

    Science.gov (United States)

    Akard, T F; Duffy, M; Hord, A; Randall, A; Sanders, A; Adelstein, K; Anani, U E; Gilmer, M J

    2018-01-01

    Legacy-making, actions or behaviors aimed at being remembered, may be one strategy to enhance coping and improve grief outcomes for bereaved parents and siblings. While legacy interventions have been developed and tested in pediatric and adult populations, legacy activities specific to bereaved parents in the neonatal intensive care unit remain unexplored. This study explored bereaved parents' perceptions of a digital storytelling legacy-making intervention for parents after the death of an infant. Six bereaved mothers and fathers participated in a focus group interview three to 12 months after the death of an infant in the NICU. A semi-structured interview guide with open-ended questions was used to obtain parent self-reports. Qualitative content analysis identified emerging themes. Four major themes emerged regarding participants' perceptions of a legacy intervention: (a) parents' willingness to participate in a legacy intervention, (b) parents' suggestions for a feasible intervention, (c) parents' suggestions for an acceptable intervention, and (d) parents' perceived benefits of legacy-making. Participants reported that a legacy-making intervention via digital storytelling would be feasible, acceptable, and beneficial for NICU parents. Study results support the need and desire for legacy-making services to be developed and offered in the NICU.

  20. Born Broken: Fonts and Information Loss in Legacy Digital Documents

    Directory of Open Access Journals (Sweden)

    Geoffrey Brown

    2011-03-01

    Full Text Available For millions of legacy documents, correct rendering depends upon resources such as fonts that are not generally embedded within the document structure. Yet there is a significant risk of information loss due to missing or incorrectly substituted fonts. Large document collections depend on thousands of unique fonts not available on a common desktop workstation, which typically has between 100 and 200 fonts. Silent substitution of fonts, performed by applications such as Microsoft Office, can yield poorly rendered documents. In this paper we use a collection of 230,000 Word documents to assess the difficulty of matching font requirements with a database of fonts. We describe the identifying information contained in common font formats, font requirements stored in Word documents, the API provided by Windows to support font requests by applications, the documented substitution algorithms used by Windows when requested fonts are not available, and the ways in which support software might be used to control font substitution in a preservation environment.

  1. Software for microcircuit systems

    International Nuclear Information System (INIS)

    Kunz, P.F.

    1978-01-01

    Modern Large Scale Integration (LSI) microcircuits are meant to be programmed in order to control the function that they perform. In the previous paper the author has already discussed the basics of microprogramming and have studied in some detail two types of new microcircuits. In this paper, methods of developing software for these microcircuits are explored. This generally requires a package of support software in order to assemble the microprogram, and also some amount of support software to test the microprograms and to test the microprogrammed circuit itself. (Auth.)

  2. Qualification of safety-critical software for digital reactor safety system in nuclear power plants

    International Nuclear Information System (INIS)

    Kwon, Kee-Choon; Park, Gee-Yong; Kim, Jang-Yeol; Lee, Jang-Soo

    2013-01-01

    This paper describes the software qualification activities for the safety-critical software of the digital reactor safety system in nuclear power plants. The main activities of the software qualification processes are the preparation of software planning documentations, verification and validation (V and V) of the software requirements specifications (SRS), software design specifications (SDS) and codes, and the testing of the integrated software and integrated system. Moreover, the software safety analysis and software configuration management are involved in the software qualification processes. The V and V procedure for SRS and SDS contains a technical evaluation, licensing suitability evaluation, inspection and traceability analysis, formal verification, software safety analysis, and an evaluation of the software configuration management. The V and V processes for the code are a traceability analysis, source code inspection, test case and test procedure generation. Testing is the major V and V activity of the software integration and system integration phases. The software safety analysis employs a hazard operability method and software fault tree analysis. The software configuration management in each software life cycle is performed by the use of a nuclear software configuration management tool. Through these activities, we can achieve the functionality, performance, reliability, and safety that are the major V and V objectives of the safety-critical software in nuclear power plants. (author)

  3. Qt based control system software for Low Energy Accelerator Facility

    International Nuclear Information System (INIS)

    Basu, A.; Singh, S.; Nagraju, S.B.V.; Gupta, S.; Singh, P.

    2012-01-01

    Qt based control system software for Low Energy Accelerating Facility (LEAF) is operational at Bhabha Atomic Research Centre (BARC), Trombay, Mumbai. LEAF is a 50 keV negative ion electrostatic accelerator based on SNICS ion source. Control system uses Nokia Trolltech's QT 4.x API for control system software. Ni 6008 USB based multifunction cards has been used for control and read back field equipments such as power supplies, pumps, valves etc. Control system architecture is designed to be client server. Qt is chosen for its excellent GUI capability and platform independent nature. Control system follows client server architecture. The paper will describe the control system. (author)

  4. A Balance of Primary and Secondary Values: Exploring a Digital Legacy

    Directory of Open Access Journals (Sweden)

    Amber L. Cushing

    2014-06-01

    Full Text Available This exploratory research explores the concept of a digital legacy as a general concept and as a collection of digital possessions with unique characteristics. The results reported in this article are part of a larger study. In Cushing (2013, the author identified the characteristics of a digital possession. In this study, these characteristics of a digital possession were utilized to explore how the characteristics of several digital possessions could form a collection, or a digital legacy. In addition to being explored as a collection of digital possessions, data was collected about the general concept of a digital legacy. In part I of the study, 23 participants from three age groups were interviewed about their general concept of a digital legacy. Five general characteristics describing a digital legacy were identified. In part II of the study, interview data from Cushing (2013 was used to create statements describing digital possessions. The statements were classified utilizing the archival concept of primary and secondary values, as well as the consumer behavior concepts of self extension to possessions and possession attachment. Primary value refers to the purpose for which the item was created, while secondary value refers to an additional value that the participants can perceive the item to hold, such as a perception that an item can represent one's identity. Using standard Q method procedure, 48 participants were directed to rank their agreement with 60 statements (written on cards, along a distribution of -5 to +5, according to the characteristics of the digital possession they would most like to maintain for a digital legacy. The ranked statements were analyzed using Q factor analysis, in order to perceive the most common statements associated with maintaining digital possessions for a digital legacy. Q method results suggested that most individuals described the digital possessions they wanted to maintain for a digital legacy using

  5. Spent nuclear fuel application of CORE reg-sign systems engineering software

    International Nuclear Information System (INIS)

    Grimm, R.J.

    1996-01-01

    The DOE has adopted a systems engineering approach for the successful completion of the Spent Nuclear Fuel (SNF) Program mission. The DOE has utilized systems engineering principles to develop the SNF program guidance documents and has held several systems engineering workshops to develop the functional hierarchies of both the programmatic and technical side of the SNF program. The sheer size and complexity of the SNF program has led to problems that the Westinghouse Savannah River Company (WSRC) is working to manage through the use of systems engineering software. WSRC began using CORE reg-sign, an off the shelf PC based software package, to assist DOE in management of the SNF program. This paper details the successful use of the CORE reg-sign systems engineering software to date and the proposed future activities

  6. A legacy building model for holistic nursing.

    Science.gov (United States)

    Lange, Bernadette; Zahourek, Rothlyn P; Mariano, Carla

    2014-06-01

    This pilot project was an effort to record the historical roots, development, and legacy of holistic nursing through the visionary spirit of four older American Holistic Nurses Association (AHNA) members. The aim was twofold: (a) to capture the holistic nursing career experiences of elder AHNA members and (b) to begin to create a Legacy Building Model for Holistic Nursing. The narratives will help initiate an ongoing, systematic method for the collection of historical data and serve as a perpetual archive of knowledge and inspiration for present and future holistic nurses. An aesthetic inquiry approach was used to conduct in-depth interviews with four older AHNA members who have made significant contributions to holistic nursing. The narratives provide a rich description of their personal and professional evolution as holistic nurses. The narratives are presented in an aesthetic format of the art forms of snapshot, pastiche, and collage rather than traditional presentations of research findings. A synopsis of the narratives is a dialogue between the three authors and provides insight for how a Legacy Model can guide our future. Considerations for practice, education, and research are discussed based on the words of wisdom from the four older holistic nurses.

  7. Research and Development on Food Nutrition Statistical Analysis Software System

    OpenAIRE

    Du Li; Ke Yun

    2013-01-01

    Designing and developing a set of food nutrition component statistical analysis software can realize the automation of nutrition calculation, improve the nutrition processional professional’s working efficiency and achieve the informatization of the nutrition propaganda and education. In the software development process, the software engineering method and database technology are used to calculate the human daily nutritional intake and the intelligent system is used to evaluate the user’s hea...

  8. Integrated conception of hardware/software mixed systems used in nuclear instrumentation

    International Nuclear Information System (INIS)

    Dias, Ailton F.; Sorel, Yves; Akil, Mohamed

    2002-01-01

    Hardware/software codesign carries out the design of systems composed by a hardware portion, with specific components, and a software portion, with microprocessor based architecture. This paper describes the Algorithm Architecture Adequation (AAA) design methodology - originally oriented to programmable multicomponent architectures, its extension to reconfigurable circuits and its application to design and development of nuclear instrumentation systems composed by programmable and configurable circuits. AAA methodology uses an unified model to describe algorithm, architecture and implementation, based on graph theory. The great advantage of AAA methodology is the utilization of a same model from the specification to the implementation of hardware/software systems, reducing the complexity and design time. (author)

  9. Embedded and real time system development a software engineering perspective concepts, methods and principles

    CERN Document Server

    Saeed, Saqib; Darwish, Ashraf; Abraham, Ajith

    2014-01-01

    Nowadays embedded and real-time systems contain complex software. The complexity of embedded systems is increasing, and the amount and variety of software in the embedded products are growing. This creates a big challenge for embedded and real-time software development processes and there is a need to develop separate metrics and benchmarks. “Embedded and Real Time System Development: A Software Engineering Perspective: Concepts, Methods and Principles” presents practical as well as conceptual knowledge of the latest tools, techniques and methodologies of embedded software engineering and real-time systems. Each chapter includes an in-depth investigation regarding the actual or potential role of software engineering tools in the context of the embedded system and real-time system. The book presents state-of-the art and future perspectives with industry experts, researchers, and academicians sharing ideas and experiences including surrounding frontier technologies, breakthroughs, innovative solutions and...

  10. Institute of legacy in the testament

    Directory of Open Access Journals (Sweden)

    MSc. Shpresa Ibrahimi

    2011-06-01

    Full Text Available Globalization as the new world order has brought to a more planned human life. This planning not only entails the individual life, but it must plan for a longer term future as well. When we talk about long terms, we immediately think about analytical skills of Roman lawyers in creating the mortis causa institute (effec-ting upon death. A characteristic of this paper comes with the latin term “leg”. The testament is a statement of will, which defines the heirs and the inheritance. While the Testament is a rather more elaborated work, the Legacy is a special provision, an order in the testament, addressed to the heirs, to submit an item or a material value to the privileged persons, called the Legatar. The Legatar, as the benefi-ciary of this provision is only a beneficiary, and does not take res-ponsibility for the debts of the inherited property. Planning of wealth may serve various functions or purposes. The Legacy represents a balance between the freedom of disposing inheritance in a free manner, and limitation of a part called nece-ssary fortune. The money or the values we decide to give away with the Institute of Legacy are not about their material value, but the significance of their investment, the goal and the best reminis-cence of the testators’ contribution in generations.

  11. Development of a methodology for assessing the safety of embedded software systems

    Science.gov (United States)

    Garrett, C. J.; Guarro, S. B.; Apostolakis, G. E.

    1993-01-01

    A Dynamic Flowgraph Methodology (DFM) based on an integrated approach to modeling and analyzing the behavior of software-driven embedded systems for assessing and verifying reliability and safety is discussed. DFM is based on an extension of the Logic Flowgraph Methodology to incorporate state transition models. System models which express the logic of the system in terms of causal relationships between physical variables and temporal characteristics of software modules are analyzed to determine how a certain state can be reached. This is done by developing timed fault trees which take the form of logical combinations of static trees relating the system parameters at different point in time. The resulting information concerning the hardware and software states can be used to eliminate unsafe execution paths and identify testing criteria for safety critical software functions.

  12. Westinghouse integrated protection system. An overview of the software design and maintenance features

    International Nuclear Information System (INIS)

    Gibson, R.J.

    1995-01-01

    The Westinghouse Integrated Protection System was designed with the goal of providing a system which can be easily verified, validated, and maintained. The software design and structure promote the ease of translation from functional requirements to applications function software while also improving the ability to verify and maintain the applications function software. The use of independent, reusable, common functions software modules focuses the design, verification, and validation of the software and reduces the likelihood of errors occurring during the application and maintenance of the software. The simple continuous loop method of operation used throughout the IPS provides a standard deterministic method of operation. The IPS design also incorporates the use of embedded self-diagnostics to perform continuous hardware oriented tests of the system and the use of an independent subsystem to automatically perform a functional test of the system. Maintenance interfaces also exist to readily identify and locate faults as well as providing other maintenance capabilities. These testing and maintenance features enhance the overall reliability and availability of the system. (orig.) (2 refs., 2 figs.)

  13. An Automated Weather Research and Forecasting (WRF)-Based Nowcasting System: Software Description

    Science.gov (United States)

    2013-10-01

    14. ABSTRACT A Web service /Web interface software package has been engineered to address the need for an automated means to run the Weather Research...An Automated Weather Research and Forecasting (WRF)- Based Nowcasting System: Software Description by Stephen F. Kirby, Brian P. Reen, and...Based Nowcasting System: Software Description Stephen F. Kirby, Brian P. Reen, and Robert E. Dumais Jr. Computational and Information Sciences

  14. The ITER Fast Plant System Controller ATCA prototype Real-Time Software Architecture

    International Nuclear Information System (INIS)

    Carvalho, B.B.; Santos, B.; Carvalho, P.F.; Neto, A.; Boncagni, L.; Batista, A.J.N.; Correia, M.; Sousa, J.; Gonçalves, B.

    2013-01-01

    Highlights: ► High performance ATCA systems for fast control and data acquisition. ► IEEE1588 timing system and synchronization. ► Plasma control algorithms. ► Real-time control software frameworks. ► Targeted for nuclear fusion experiments with long duration discharges. -- Abstract: IPFN is developing a prototype Fast Plant System Controller (FPSC) based in ATCA embedded technologies dedicated to ITER CODAC data acquisition and control tasks in the sub-millisecond range. The main goal is to demonstrate the usability of the ATCA standard and its enhanced specifications for the high speed, very high density parallel data acquisition needs of the most demanding ITER tokamak plasma Instrumentation and Control (I and C) systems. This effort included the in-house development of a new family of high performance ATCA I/O and timing boards. The standard ITER software system CODAC Core System (CCS) v3.1, with the control based in the EPICS system does not cover yet the real-time requirements fulfilled by this hardware, so a new set of software components was developed for this specific platform, attempting to integrate and leverage the new features in CSS, for example the Multithreaded Application Real Time executor (MARTe) software framework, the new Data Archiving Network (DAN) solution, an ATCA IEEE-1588-2008 timing interface, and the Intelligent Platform Management Interface (IPMI) for system monitoring and remote management. This paper presents the overall software architecture for the ATCA FPSC, as well a discussion on the ITER constrains and design choices and finally a detailed description of the software components already developed

  15. The ITER Fast Plant System Controller ATCA prototype Real-Time Software Architecture

    Energy Technology Data Exchange (ETDEWEB)

    Carvalho, B.B., E-mail: bernardo@ipfn.ist.utl.pt [Associacao EURATOM/IST Instituto de Plasmas e Fusao Nuclear, Instituto Superior Tecnico, Universidade Tecnica de Lisboa, P-1049-001 Lisboa (Portugal); Santos, B.; Carvalho, P.F.; Neto, A. [Associacao EURATOM/IST Instituto de Plasmas e Fusao Nuclear, Instituto Superior Tecnico, Universidade Tecnica de Lisboa, P-1049-001 Lisboa (Portugal); Boncagni, L. [Associazione Euratom-ENEA sulla Fusione, Frascati Research Centre, Division of Fusion Physics, Frascati, Rome (Italy); Batista, A.J.N.; Correia, M.; Sousa, J.; Gonçalves, B. [Associacao EURATOM/IST Instituto de Plasmas e Fusao Nuclear, Instituto Superior Tecnico, Universidade Tecnica de Lisboa, P-1049-001 Lisboa (Portugal)

    2013-10-15

    Highlights: ► High performance ATCA systems for fast control and data acquisition. ► IEEE1588 timing system and synchronization. ► Plasma control algorithms. ► Real-time control software frameworks. ► Targeted for nuclear fusion experiments with long duration discharges. -- Abstract: IPFN is developing a prototype Fast Plant System Controller (FPSC) based in ATCA embedded technologies dedicated to ITER CODAC data acquisition and control tasks in the sub-millisecond range. The main goal is to demonstrate the usability of the ATCA standard and its enhanced specifications for the high speed, very high density parallel data acquisition needs of the most demanding ITER tokamak plasma Instrumentation and Control (I and C) systems. This effort included the in-house development of a new family of high performance ATCA I/O and timing boards. The standard ITER software system CODAC Core System (CCS) v3.1, with the control based in the EPICS system does not cover yet the real-time requirements fulfilled by this hardware, so a new set of software components was developed for this specific platform, attempting to integrate and leverage the new features in CSS, for example the Multithreaded Application Real Time executor (MARTe) software framework, the new Data Archiving Network (DAN) solution, an ATCA IEEE-1588-2008 timing interface, and the Intelligent Platform Management Interface (IPMI) for system monitoring and remote management. This paper presents the overall software architecture for the ATCA FPSC, as well a discussion on the ITER constrains and design choices and finally a detailed description of the software components already developed.

  16. Software simulation: a tool for enhancing control system design

    International Nuclear Information System (INIS)

    Sze, B.; Ridgway, G.H.

    2008-01-01

    The creation, implementation and management of engineering design tools are important to the quality and efficiency of any large engineering project. Some of the most complicated tools to develop are system simulators. The development and implementation of system simulators to support replacement fuel handling control systems is of particular interest to the Canadian nuclear industry given the current age of installations and the risk of obsolescence to many utilities. The use of such simulator tools has been known to significantly improve successful deployment of new software packages and maintenance-related software changes while reducing the time required for their overall development. Moreover, these simulation systems can also serve as operator training stations and provide a virtual environment for site engineers to test operational changes before they are uploaded to the actual system. (author)

  17. Stakeholder reflections of the tourism and nation-branding legacy of ...

    African Journals Online (AJOL)

    Over the past decade there has been a growing awareness of the significant impact that hosting mega sport events can have on a nation's brand. This paper discusses the context of nation-branding as a tourism legacy and the role of mega sport events in generating a tourism and nation-branding legacy in relation to the ...

  18. The ATLAS software installation system for LCG/EGEE

    Energy Technology Data Exchange (ETDEWEB)

    Salvo, A D [Istituto Nazionale di Fisica Nucleare, sez. Roma 1 (Italy); Barchiesi, A [Universita di Roma I ' La Sapienza' (Italy); Gnanvo, K [Queen Mary and Westfield College (United Kingdom); Gwilliam, C [University of Liverpool (United Kingdom); Kennedy, J; Krobath, G [Ludwig-Maximilians-Universitaet Muenchen (Germany); Olszewski, A [Henryk Niewodniczanski Institute of Nuclear Physics, Polish Academy of Sciences (Poland); Rybkine, G [Royal Holloway College (United Kingdom)

    2008-07-15

    The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within a few hours, have driven the need for an automatic installation system for the LHC experiments. In this work we describe the ATLAS system for the experiment software installation in LCG/EGEE, based on the Light Job Submission Framework for Installation (LJSFi), an independent job submission framework for generic submission and job tracking in EGEE. LJSFi is able to automatically discover, check, install, test and tag the full set of resources made available in LCG/EGEE to the ATLAS Virtual Organization in a few hours, depending on the site availability.

  19. The design and performance of a scintillating-fibre tracker for the cosmic-ray muon tomography of legacy nuclear waste containers

    Science.gov (United States)

    Clarkson, A.; Hamilton, D. J.; Hoek, M.; Ireland, D. G.; Johnstone, J. R.; Kaiser, R.; Keri, T.; Lumsden, S.; Mahon, D. F.; McKinnon, B.; Murray, M.; Nutbeam-Tuffs, S.; Shearer, C.; Staines, C.; Yang, G.; Zimmerman, C.

    2014-05-01

    Tomographic imaging techniques using the Coulomb scattering of cosmic-ray muons are increasingly being exploited for the non-destructive assay of shielded containers in a wide range of applications. One such application is the characterisation of legacy nuclear waste materials stored within industrial containers. The design, assembly and performance of a prototype muon tomography system developed for this purpose are detailed in this work. This muon tracker comprises four detection modules, each containing orthogonal layers of Saint-Gobain BCF-10 2 mm-pitch plastic scintillating fibres. Identification of the two struck fibres per module allows the reconstruction of a space point, and subsequently, the incoming and Coulomb-scattered muon trajectories. These allow the container content, with respect to the atomic number Z of the scattering material, to be determined through reconstruction of the scattering location and magnitude. On each detection layer, the light emitted by the fibre is detected by a single Hamamatsu H8500 MAPMT with two fibres coupled to each pixel via dedicated pairing schemes developed to ensure the identification of the struck fibre. The PMT signals are read out to standard charge-to-digital converters and interpreted via custom data acquisition and analysis software. The design and assembly of the detector system are detailed and presented alongside results from performance studies with data collected after construction. These results reveal high stability during extended collection periods with detection efficiencies in the region of 80% per layer. Minor misalignments of millimetre order have been identified and corrected in software. A first image reconstructed from a test configuration of materials has been obtained using software based on the Maximum Likelihood Expectation Maximisation algorithm. The results highlight the high spatial resolution provided by the detector system. Clear discrimination between the low, medium and high

  20. Systematic development of industrial control systems using Software/Hardware Engineering

    NARCIS (Netherlands)

    Voeten, J.P.M.; van der Putten, P.H.A.; Stevens, M.P.J.; Milligan, P.; Corr, P.

    1997-01-01

    SHE (Software/Hardware Engineering) is a new object-oriented analysis, specification and design method for complex reactive hardware/software systems. SHE is based on the formal specification language POOSL and a design framework guiding analysis and design activities. This paper reports on the

  1. Paralympic Games: History and Legacy of a Global Movement.

    Science.gov (United States)

    Legg, David

    2018-05-01

    The Paralympic Games have an interesting history that began after World War II. The Games and movement have been impacted by and have had an impact on society and the larger able-bodied sport system. The future of the Games and movement is also further impacted by larger cultural shifts, and the Games themselves have potentially left lasting legacies for the host cities and persons with impairment worldwide. Copyright © 2018 Elsevier Inc. All rights reserved.

  2. Software engineering for the EBR-II data acquisition system conversion

    International Nuclear Information System (INIS)

    Schorzman, W.

    1988-01-01

    The original data acquisition system (DAS) for the Experimental Breeder Reactor II (EBR-II) was placed into service with state-of-the-art computer and peripherals in 1970. Software engineering principles for real-time data acquisition were in their infancy, and the original software design was dictated by limited hardware resources. The functional requirements evolved from creative ways to gather and display data. This abstract concept developed into an invaluable tool for system analysis, data reporting, and as a plant monitor for operations. In this paper the approach is outlined to the software conversion project with the restraints of operational transparency and 6 weeks for final conversion and testing. The outline is then compared with the formal principles of software engineering to show the way that bridge the gap can be bridged between the theoretical and real world by analyzing the work and listing the lessons learned

  3. Differentiating drought legacy effects on vegetation growth over the temperate Northern Hemisphere

    Czech Academy of Sciences Publication Activity Database

    Wu, X.; Liu, H.; Li, X.; Ciais, P.; Babst, F.; Guo, W.; Zhang, C.; Magliulo, V.; Pavelka, Marian; Liu, S.; Huang, Y.; Wang, P.; Shi, C. D.; Ma, Y.

    2018-01-01

    Roč. 24, č. 1 (2018), s. 504-516 ISSN 1354-1013 EU Projects: European Commission(XE) 640176 Institutional support: RVO:86652079 Keywords : drought legacy effect * drought resilience * ecohydrological responses * extreme drought * plant functional groups * rooting system * stomatal conductance * vegetation growth Subject RIV: EH - Ecology, Behaviour Impact factor: 8.502, year: 2016

  4. An effective technique for the software requirements analysis of NPP safety-critical systems, based on software inspection, requirements traceability, and formal specification

    International Nuclear Information System (INIS)

    Koo, Seo Ryong; Seong, Poong Hyun; Yoo, Junbeom; Cha, Sung Deok; Yoo, Yeong Jae

    2005-01-01

    A thorough requirements analysis is indispensable for developing and implementing safety-critical software systems such as nuclear power plant (NPP) software systems because a single error in the requirements can generate serious software faults. However, it is very difficult to completely analyze system requirements. In this paper, an effective technique for the software requirements analysis is suggested. For requirements verification and validation (V and V) tasks, our technique uses software inspection, requirement traceability, and formal specification with structural decomposition. Software inspection and requirements traceability analysis are widely considered the most effective software V and V methods. Although formal methods are also considered an effective V and V activity, they are difficult to use properly in the nuclear fields as well as in other fields because of their mathematical nature. In this work, we propose an integrated environment (IE) approach for requirements, which is an integrated approach that enables easy inspection by combining requirement traceability and effective use of a formal method. The paper also introduces computer-aided tools for supporting IE approach for requirements. Called the nuclear software inspection support and requirements traceability (NuSISRT), the tool incorporates software inspection, requirement traceability, and formal specification capabilities. We designed the NuSISRT to partially automate software inspection and analysis of requirement traceability. In addition, for the formal specification and analysis, we used the formal requirements specification and analysis tool for nuclear engineering (NuSRS)

  5. A software application for energy flow simulation of a grid connected photovoltaic system

    International Nuclear Information System (INIS)

    Hamad, Ayman A.; Alsaad, Mohammad A.

    2010-01-01

    A computer software application was developed to simulate hourly energy flow of a grid connected photovoltaic system. This software application enables conducting an operational evaluation of a studied photovoltaic system in terms of energy exchange with the electrical grid. The system model consists of a photovoltaic array, a converter and an optional generic energy storage component that supports scheduled charging/discharging. In addition to system design parameters, the software uses hourly solar data and hourly load data to determine the amount of energy exchanged with electrical grid for each hour of the simulated year. The resulting information is useful in assessing the impact of the system on demand for electrical energy of a building that uses it. The software also aggregates these hourly results in daily, monthly and full year sums. The software finds the financial benefit of the system as the difference in grid electrical energy cost between two simultaneously considered cases. One is with load supplied only by the electrical grid, while the other is with the photovoltaic system present and contributing energy. The software supports the energy pricing scheme used in Jordan for domestic consumers, which is based on slices of monthly consumption. By projecting the yearly financial results on the system lifetime, the application weighs the financial benefit resulting from using the system against its cost, thus facilitating an economical evaluation.

  6. Substantially Evolutionary Theorizing in Designing Software-Intensive Systems

    Directory of Open Access Journals (Sweden)

    Petr Sosnin

    2018-04-01

    Full Text Available Useful inheritances from scientific experience open perspective ways for increasing the degree of success in designing of systems with software. One such way is a search and build applied theory that takes into account the nature of design and the specificity of software engineering. This paper presents a substantially evolutionary approach to creating the project theories, the application of which leads to positive effects that are traditionally expected from theorizing. Any implementation of the approach is based on a reflection by designers of an operational space of designing onto a semantic memory of a question-answer type. One of the results of such reflection is a system of question-answer nets, the nodes of which register facts of interactions of designers with accessible experience. A set of such facts is used by designers for creating and using the theory that belongs to the new subclass of Grounded Theories. This sub-class is oriented on organizationally behavioral features of a project’s work based on design thinking, automated mental imagination, and thought experimenting that facilitate increasing the degree of controlled intellectualization in the design process and, correspondingly, increasing the degree of success in the development of software-intensive systems.

  7. Features of commercial computer software systems for medical examiners and coroners.

    Science.gov (United States)

    Hanzlick, R L; Parrish, R G; Ing, R

    1993-12-01

    There are many ways of automating medical examiner and coroner offices, one of which is to purchase commercial software products specifically designed for death investigation. We surveyed four companies that offer such products and requested information regarding each company and its hardware, software, operating systems, peripheral devices, applications, networking options, programming language, querying capability, coding systems, prices, customer support, and number and size of offices using the product. Although the four products (CME2, ForenCIS, InQuest, and Medical Examiner's Software System) are similar in many respects and each can be installed on personal computers, there are differences among the products with regard to cost, applications, and the other features. Death investigators interested in office automation should explore these products to determine the usefulness of each in comparison with the others and in comparison with general-purpose, off-the-shelf databases and software adaptable to death investigation needs.

  8. Legacies of the Manhattan Project

    Science.gov (United States)

    Kevles, Daniel

    2017-01-01

    The Manhattan Project of World War II mobilized thousands of people, including many of the nation's leading physicists, and extensive material resources to design, develop, and manufacture the world's first nuclear weapons. It also established sprawling new facilities for the production of fissionable fuels - notably at Oak Ridge, Tennessee, and Hanford, Washington. It left a set of powerful legacies in the context of the Cold War - endowing scientists with conscience-taxing responsibilities in the nuclear arms race; promoting enormous patronage of academic research by defense and defense-related federal agencies, notably the Office of Naval Research and the Atomic Energy Commission; and turning its wartime facilities into major national laboratories that advanced the fields of high-energy and nuclear physics and stimulated local industrial economies but that in some cases, notably at Hanford, severely polluted the surrounding environment with radioactive waste and disrupted the livelihoods of native peoples. ``Legacies of the Manhattan Project''

  9. Software Tools to Support the Assessment of System Health

    Science.gov (United States)

    Melcher, Kevin J.

    2013-01-01

    This presentation provides an overview of three software tools that were developed by the NASA Glenn Research Center to support the assessment of system health: the Propulsion Diagnostic Method Evaluation Strategy (ProDIMES), the Systematic Sensor Selection Strategy (S4), and the Extended Testability Analysis (ETA) tool. Originally developed to support specific NASA projects in aeronautics and space, these software tools are currently available to U.S. citizens through the NASA Glenn Software Catalog. The ProDiMES software tool was developed to support a uniform comparison of propulsion gas path diagnostic methods. Methods published in the open literature are typically applied to dissimilar platforms with different levels of complexity. They often address different diagnostic problems and use inconsistent metrics for evaluating performance. As a result, it is difficult to perform a one ]to ]one comparison of the various diagnostic methods. ProDIMES solves this problem by serving as a theme problem to aid in propulsion gas path diagnostic technology development and evaluation. The overall goal is to provide a tool that will serve as an industry standard, and will truly facilitate the development and evaluation of significant Engine Health Management (EHM) capabilities. ProDiMES has been developed under a collaborative project of The Technical Cooperation Program (TTCP) based on feedback provided by individuals within the aircraft engine health management community. The S4 software tool provides a framework that supports the optimal selection of sensors for health management assessments. S4 is structured to accommodate user ]defined applications, diagnostic systems, search techniques, and system requirements/constraints. One or more sensor suites that maximize this performance while meeting other user ]defined system requirements that are presumed to exist. S4 provides a systematic approach for evaluating combinations of sensors to determine the set or sets of

  10. Design and implementation of embedded Bluetooth software system

    Science.gov (United States)

    Zhou, Zhijian; Zhou, Shujie; Xu, Huimin

    2001-10-01

    This thesis introduces the background knowledge and characteristics of Bluetooth technology. Then it summarizes the architecture and working principle of Bluetooth software. After carefully studying the characteristics of embedded operating system and Bluetooth software, this thesis declared two sets of module about Bluetooth software. Corresponding to these module's characteristics, this thesis introduces the design and implementation of LAN Access and Bluetooth headset. The Headset part introduces a developing method corresponding to the particularity of Bluetooth control software. Although these control software are application entity, the control signaling exchanged between them are regulations according to former definitions and they functions through the interaction of data and control information. These data and control information construct the protocol data unit (PDU), and the former definition can be seen as protocol in fact. This thesis uses the advanced development flow on communication protocol development as reference, a formal method - SDL (Specification and Description Language) - describing, validating and coding manually to C. This method not only reserved the efficiency of manually coded code, but also it ensures the quality of codes. The introduction also involves finite state machine theory while introduces the practical developing method on protocol development with the aid of SDL.

  11. Software-Based Student Response Systems: An Interdisciplinary Initiative

    Science.gov (United States)

    Fischer, Carol M.; Hoffman, Michael S.; Casey, Nancy C.; Cox, Maureen P.

    2015-01-01

    Colleagues from information technology and three academic departments collaborated on an instructional technology initiative to employ student response systems in classes in mathematics, accounting and education. The instructors assessed the viability of using software-based systems to enable students to use their own devices (cell phones,…

  12. Software Defined Common Processing System (SDCPS), Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — Coherent Logix, Incorporated (CLX) proposes the development of a Software Defined Common Processing System (SDCPS) that leverages the inherent advantages of an...

  13. The VAXONLINE software system at Fermilab

    International Nuclear Information System (INIS)

    White, V.; Heinicke, P.; Berman, E.

    1987-06-01

    The VAXONLINE software system, started in late 1984, is now in use at 12 experiments at Fermilab, with at least one VAX or MicroVax. Data acquisition features now provide for the collection and combination of data from one or more sources, via a list-driven Event Builder program. Supported sources include CAMAC, FASTBUS, Front-end PDP-11's, Disk, Tape, DECnet, and other processors running VAXONLINE. This paper describes the functionality provided by the VAXONLINE system, gives performance figures, and discusses the ongoing program of enhancements

  14. Licensing process for safety-critical software-based systems

    Energy Technology Data Exchange (ETDEWEB)

    Haapanen, P. [VTT Automation, Espoo (Finland); Korhonen, J. [VTT Electronics, Espoo (Finland); Pulkkinen, U. [VTT Automation, Espoo (Finland)

    2000-12-01

    System vendors nowadays propose software-based technology even for the most critical safety functions in nuclear power plants. Due to the nature of software faults and the way they cause system failures new methods are needed for the safety and reliability evaluation of these systems. In the research project 'Programmable automation systems in nuclear power plants (OHA)', financed together by the Radiation and Nuclear Safety Authority (STUK), the Ministry of Trade and Industry (KTM) and the Technical Research Centre of Finland (VTT), various safety assessment methods and tools for software based systems are developed and evaluated. As a part of the OHA-work a reference model for the licensing process for software-based safety automation systems is defined. The licensing process is defined as the set of interrelated activities whose purpose is to produce and assess evidence concerning the safety and reliability of the system/application to be licensed and to make the decision about the granting the construction and operation permissions based on this evidence. The parties of the licensing process are the authority, the licensee (the utility company), system vendors and their subcontractors and possible external independent assessors. The responsibility about the production of the evidence in first place lies at the licensee who in most cases rests heavily on the vendor expertise. The evaluation and gauging of the evidence is carried out by the authority (possibly using external experts), who also can acquire additional evidence by using their own (independent) methods and tools. Central issue in the licensing process is to combine the quality evidence about the system development process with the information acquired through tests, analyses and operational experience. The purpose of the licensing process described in this report is to act as a reference model both for the authority and the licensee when planning the licensing of individual applications

  15. Licensing process for safety-critical software-based systems

    International Nuclear Information System (INIS)

    Haapanen, P.; Korhonen, J.; Pulkkinen, U.

    2000-12-01

    System vendors nowadays propose software-based technology even for the most critical safety functions in nuclear power plants. Due to the nature of software faults and the way they cause system failures new methods are needed for the safety and reliability evaluation of these systems. In the research project 'Programmable automation systems in nuclear power plants (OHA)', financed together by the Radiation and Nuclear Safety Authority (STUK), the Ministry of Trade and Industry (KTM) and the Technical Research Centre of Finland (VTT), various safety assessment methods and tools for software based systems are developed and evaluated. As a part of the OHA-work a reference model for the licensing process for software-based safety automation systems is defined. The licensing process is defined as the set of interrelated activities whose purpose is to produce and assess evidence concerning the safety and reliability of the system/application to be licensed and to make the decision about the granting the construction and operation permissions based on this evidence. The parties of the licensing process are the authority, the licensee (the utility company), system vendors and their subcontractors and possible external independent assessors. The responsibility about the production of the evidence in first place lies at the licensee who in most cases rests heavily on the vendor expertise. The evaluation and gauging of the evidence is carried out by the authority (possibly using external experts), who also can acquire additional evidence by using their own (independent) methods and tools. Central issue in the licensing process is to combine the quality evidence about the system development process with the information acquired through tests, analyses and operational experience. The purpose of the licensing process described in this report is to act as a reference model both for the authority and the licensee when planning the licensing of individual applications. Many of the

  16. Spent nuclear fuel application of CORE reg-sign systems engineering software

    International Nuclear Information System (INIS)

    Grimm, R.J.

    1996-01-01

    The Department of Energy (DOE) has adopted a systems engineering approach for the successful completion of the Spent Nuclear Fuel (SNF) Program mission. The DOE has utilized systems engineering principles to develop the SNF Program guidance documents and has held several systems engineering workshops to develop the functional hierarchies of both the programmatic and technical side of the SNF Program. The sheer size and complexity of the SNF Program, however, has led to problems that the Westinghouse Savannah River Company (WSRC) is working to manage through the use of systems engineering software. WSRC began using CORE reg-sign, an off-the-shelf PC based software package, to assist the DOE in management of the SNF program. This paper details the successful use of the CORE reg-sign systems engineering software to date and the proposed future activities

  17. A systematic review of applying modern software engineering techniques to developing robotic systems

    Directory of Open Access Journals (Sweden)

    Claudia Pons

    2012-01-01

    Full Text Available Robots have become collaborators in our daily life. While robotic systems become more and more complex, the need to engineer their software development grows as well. The traditional approaches used in developing these software systems are reaching their limits; currently used methodologies and tools fall short of addressing the needs of such complex software development. Separating robotics’ knowledge from short-cycled implementation technologies is essential to foster reuse and maintenance. This paper presents a systematic review (SLR of the current use of modern software engineering techniques for developing robotic software systems and their actual automation level. The survey was aimed at summarizing existing evidence concerning applying such technologies to the field of robotic systems to identify any gaps in current research to suggest areas for further investigation and provide a background for positioning new research activities.

  18. Einstein's Legacy, at the Globe

    CERN Multimedia

    2005-01-01

    One-hundred years on, Albert Einstein's theories continue to fuel the daily work of physicists. From research into gravity waves to the quest for grand unification in physics, today's researchers have not finished with the legacy of the most famous and iconic physicist of the 20th Century.

  19. A legacy endures. A Maine system emphasizes its sponsor's mission in all aspects of its work.

    Science.gov (United States)

    Stapleton, Marguerite

    2005-01-01

    The Sisters of Charity Health System, Lewiston, ME, a member of Covenant Health Systems, Lexington, MA, remains deeply committed to the mission of service begun by its foundress, St. Marguerite d'Youville. Although St. Marguerite experienced a hard life, her resilience and her commitment to the poor and disadvantaged serve as an inspiration to those who continue her legacy of compassionate care. The founding work of St. Marguerite and the sisters has helped to foster a culture in which the mission of service thrives among the system's 2,000 employees. This culture can be attributed to two things: the system's organizational values of compassion, stewardship, respect, and excellence; and the recognition of those employees whose work embodies these values. From the boardroom to the patient room, mission is integrated into each decision and action. Every two years, each of Covenant Health System's member facilities engages in a mission assessment process that examines various aspects of mission, including Catholic identity, holistic care, care for the poor, mission values integration, ethics and employee relations. In addition, the Sisters of Charity Health System's board has its own standing Mission and Community Committee, which looks strategically at how creatively and faithfully the system is continuing to live its mission.

  20. The software design of area γ radiation monitoring system

    International Nuclear Information System (INIS)

    Song Chenxin; Deng Changming; Cheng Chang; Ren Yi; Meng Dan; Liu Yun

    2007-01-01

    This paper main introduction the system structure, software architecture, design ideas of the area γ radiation monitoring system. Detailed introduction some programming technology about the computer communication with the local display unit. (authors)