WorldWideScience

Sample records for unix process control

  1. Implementation of a control system test environment in UNIX

    International Nuclear Information System (INIS)

    Brittain, C.R.; Otaduy, P.J.; Rovere, L.A.

    1990-01-01

    This paper discusses how UNIX features such as shared memory, remote procedure calls, and signalling have been used to implement a distributed computational environment ideal for the development and testing of digital control systems. The resulting environment -based on features commonly available in commercial workstations- is flexible, allows process simulation and controller development to proceed in parallel, and provides for testing and validation in a realistic environment. In addition, the use of shared memory to exchange data allows other tasks such as user interfaces and recorders to be added without affecting the process simulation or controllers. A library of functions is presented which provides a simple interface to using the features described. These functions can be used in either C or FORTRAN programs and have been tested on a network of Sun workstations and an ENCORE parallel computer. 6 refs., 2 figs

  2. The hybrid UNIX controller for real-time data acquisition

    International Nuclear Information System (INIS)

    Huesman, R.H.; Klein, G.J.; Fleming, T.K.

    1996-01-01

    The authors describe a hybrid data acquisition architecture integrating a conventional UNIX workstation with CAMAC-based real-time hardware. The system combines the high-level programming simplicity and user interface of a UNIX workstation with the low-level timing control available from conventional real-time hardware. They detail this architecture as it has been implemented for control of the Donner 600-Crystal Positron Tomograph (PET600). Low-level data acquisition is carried out in this system using eight LeCroy 3588 histogrammers, which together after derandomization, acquire events at rates up to 4 MHz, and two dedicated Motorola 6809 microprocessors, which arbitrate fine timing control during acquisition. A SUN Microsystems UNIX workstation is used for high-level control, allowing an easily extensible user interface in an X-Windows environment, as well as real-time communications to the low-level acquisition units. Communication between the high- and low-level units is carried out via a Jorway 73A SCSI-CAMAC crate controller and a serial interface. For this application, the hybrid configuration segments low from high-level control for ease of maintenance and provided a low-cost upgrade from dated high-level control hardware

  3. Utilities in UNIX; Programas de Utilidad UNIX

    Energy Technology Data Exchange (ETDEWEB)

    Perez, L

    2002-07-01

    This manual goes to the users with some or much experience in the unix operating system. In such manner that they can get more efficiency using the unix of the most vendors. Include the majority of UNIX commands, shell built-in functions to create scripts, and a brief explication of the variables in several environments. In addition, other products are included, more and more integrated in the most of the unix operating systems. For example; the scanning and processing language awk, the print server LPRng, GNU utilities, batch subsystem, etc. The manual was initially based in an specific unix. But it had been written for use of the most unix that exist: Tru64 unix, aix, iris, hpux, solaris y linux. In this way, many examples in the chapter had been included. The purpose of this manual is to provide an UNIX reference for advanced users in any of the unix operating systems family. (Author)

  4. Utilities in UNIX

    International Nuclear Information System (INIS)

    Perez, L.

    2002-01-01

    This manual goes to the users with some or much experience in the unix operating system. In such manner that they can get more efficiency using the unix of the most vendors. Include the majority of UNIX commands, shell built-in functions to create scripts, and a brief explication of the variables in several environments. In addition, other products are included, more and more integrated in the most of the unix operating systems. For example: the scanning and processing language awk, the print server LPRng, GNU Utilities, batch subsystem, etc. The manual was initially based in an specific unix. But it and been written for use of the most unix that exist: Tru64 unix, aix, iris, hpux. solaris y linux. In this way, many examples in the chapter had been included. The purpose of this manual is to provide an UNIX reference for advanced users in any of the unix operating systems family. (Author)

  5. Progress of data processing system in JT-60 utilizing the UNIX-based workstations

    International Nuclear Information System (INIS)

    Sakata, Shinya; Kiyono, Kimihiro; Oshima, Takayuki; Sato, Minoru; Ozeki, Takahisa

    2007-07-01

    JT-60 data processing system (DPS) possesses three-level hierarchy. At the top level of hierarchy is JT-60 inter-shot processor (MSP-ISP), which is a mainframe computer, provides communication with the JT-60 supervisory control system and supervises the internal communication inside the DPS. The middle level of hierarchy has minicomputers and the bottom level of hierarchy has individual diagnostic subsystems, which consist of the CAMAC and VME modules. To meet the demand for advanced diagnostics, the DPS has been progressed in stages from a three-level hierarchy system, which was dependent on the processing power of the MSP-ISP, to a two-level hierarchy system, which is decentralized data processing system (New-DPS) by utilizing the UNIX-based workstations and network technology. This replacement had been accomplished, and the New-DPS has been started to operate in October 2005. In this report, we describe the development and improvement of the New-DPS, whose functions were decentralized from the MSP-ISP to the UNIX-based workstations. (author)

  6. The equipment access software for a distributed UNIX-based accelerator control system

    International Nuclear Information System (INIS)

    Trofimov, Nikolai; Zelepoukine, Serguei; Zharkov, Eugeny; Charrue, Pierre; Gareyte, Claire; Poirier, Herve

    1994-01-01

    This paper presents a generic equipment access software package for a distributed control system using computers with UNIX or UNIX-like operating systems. The package consists of three main components, an application Equipment Access Library, Message Handler and Equipment Data Base. An application task, which may run in any computer in the network, sends requests to access equipment through Equipment Library calls. The basic request is in the form Equipment-Action-Data and is routed via a remote procedure call to the computer to which the given equipment is connected. In this computer the request is received by the Message Handler. According to the type of the equipment connection, the Message Handler either passes the request to the specific process software in the same computer or forwards it to a lower level network of equipment controllers using MIL1553B, GPIB, RS232 or BITBUS communication. The answer is then returned to the calling application. Descriptive information required for request routing and processing is stored in the real-time Equipment Data Base. The package has been written to be portable and is currently available on DEC Ultrix, LynxOS, HPUX, XENIX, OS-9 and Apollo domain. ((orig.))

  7. Unix Philosophy and the Real World: Control Software for Humanoid Robots

    Directory of Open Access Journals (Sweden)

    Neil Thomas Dantam

    2016-03-01

    Full Text Available Robot software combines the challenges of general purpose and real-time software, requiring complex logic and bounded resource use. Physical safety, particularly for dynamic systems such as humanoid robots, depends on correct software. General purpose computation has converged on unix-like operating systems -- standardized as POSIX, the Portable Operating System Interface -- for devices from cellular phones to supercomputers. The modular, multi-process design typical of POSIX applications is effective for building complex and reliable software. Absent from POSIX, however, is an interproccess communication mechanism that prioritizes newer data as typically desired for control of physical systems. We address this need in the Ach communication library which provides suitable semantics and performance for real-time robot control. Although initially designed for humanoid robots, Ach has broader applicability to complex mechatronic devices -- humanoid and otherwise -- that require real-time coupling of sensors, control, planning, and actuation. The initial user space implementation of Ach was limited in the ability to receive data from multiple sources. We remove this limitation by implementing Ach as a Linux kernel module, enabling Ach's high-performance and latest-message-favored semantics within conventional POSIX communication pipelines. We discuss how these POSIX interfaces and design principles apply to robot software, and we present a case study using the Ach kernel module for communication on the Baxter robot.

  8. UNIX by examples

    International Nuclear Information System (INIS)

    Nee, F.

    1990-10-01

    This report discusses the following topics on UNIX basis programming: file structure; frequently used commands/utilities; control structures used in shell script; c-shell programming; and bourne shell programming

  9. UNIX code management and distribution

    International Nuclear Information System (INIS)

    Hung, T.; Kunz, P.F.

    1992-09-01

    We describe a code management and distribution system based on tools freely available for the UNIX systems. At the master site, version control is managed with CVS, which is a layer on top of RCS, and distribution is done via NFS mounted file systems. At remote sites, small modifications to CVS provide for interactive transactions with the CVS system at the master site such that remote developers are true peers in the code development process

  10. Development of a new discharge control system utilizing UNIX workstations and VME-bus systems for JT-60

    Energy Technology Data Exchange (ETDEWEB)

    Akasaka, Hiromi; Sueoka, Michiharu; Takano, Shoji; Totsuka, Toshiyuki; Yonekawa, Izuru; Kurihara, Kenichi; Kimura, Toyoaki [Japan Atomic Energy Research Inst., Naka, Ibaraki (Japan). Naka Fusion Research Establishment

    2002-01-01

    The JT-60 discharge control system, which had used HIDIC-80E 16 bit mini-computers and CAMAC systems since the start of JT-60 experiment in 1985, was renewed in March, 2001. The new system consists of a UNIX workstation and a VME-bus system, and features a distributed control system. The workstation performs message communication with a VME-bus system and controllers of JT-60 sub-systems and processing for discharge control because of its flexibility to construction of a new network and modifications of software. The VME-bus system performs discharge sequence control because it is suitable for fast real time control and flexible to the hardware extension. The replacement has improved the control function and reliability of the discharge control system and also has provided sufficient performance necessary for future modifications of JT-60. The new system has been running successfully since April 2001. The data acquisition speed was confirmed to be twice faster than the previous one. This report describes major functions of the discharge control system, technical ideas for developing the system and results of the initial operation in detail. (author)

  11. Fermi UNIX trademark environment

    International Nuclear Information System (INIS)

    Nicholls, J.

    1991-03-01

    The introduction of UNIX at Fermilab involves multiple platforms and multiple vendors. Additionally, a single user may have to use more than one platform. This heterogeneity and multiplicity makes it necessary to define a Fermilab environment for UNIX so that as much as possible the systems ''look and feel'' the same. We describe our environment, including both the commercial products and the local tools used to support it. Other products designed for the UNIX environment are also described. 19 refs

  12. A unix configuration engine

    International Nuclear Information System (INIS)

    Burgess, M.

    1994-06-01

    A high level description language is presented for the purpose of automatically configuring large heterogeneous networked unix environments, based on class-oriented abstractions. The configuration engine is portable and easily extensible

  13. Unix Application Migration Guide

    CERN Document Server

    Microsoft. Redmond

    2003-01-01

    Drawing on the experience of Microsoft consultants working in the field, as well as external organizations that have migrated from UNIX to Microsoft® Windows®, this guide offers practical, prescriptive guidance on the issues you are likely to face when porting existing UNIX applications to the Windows operating system environment. Senior IT decision makers, network managers, and operations managers will get real-world guidance and best practices on planning and implementation issues to understand the different methods through which migration or co-existence can be accomplished. Also detailing

  14. Unix Security Cookbook

    Science.gov (United States)

    Rehan, S. C.

    This document has been written to help Site Managers secure their Unix hosts from being compromised by hackers. I have given brief introductions to the security tools along with downloading, configuring and running information. I have also included a section on my recommendations for installing these security tools starting from an absolute minimum security requirement.

  15. Development of a UNIX network compatible reactivity computer

    International Nuclear Information System (INIS)

    Sanchez, R.F.; Edwards, R.M.

    1996-01-01

    A state-of-the-art UNIX network compatible controller and UNIX host workstation with MATLAB/SIMULINK software were used to develop, implement, and validate a digital reactivity calculation. An objective of the development was to determine why a Macintosh-based reactivity computer reactivity output drifted intolerably

  16. Doing accelerator physics using SDDS, UNIX, and EPICS

    International Nuclear Information System (INIS)

    Borland, M.; Emery, L.; Sereno, N.

    1995-01-01

    The use of the SDDS (Self-Describing Data Sets) file protocol, together with the UNIX operating system and EPICS (Experimental Physics and Industrial Controls System), has proved powerful during the commissioning of the APS (Advanced Photon Source) accelerator complex. The SDDS file protocol has permitted a tool-oriented approach to developing applications, wherein generic programs axe written that function as part of multiple applications. While EPICS-specific tools were written for data collection, automated experiment execution, closed-loop control, and so forth, data processing and display axe done with the SDDS Toolkit. Experiments and data reduction axe implemented as UNIX shell scripts that coordinate the execution of EPICS specific tools and SDDS tools. Because of the power and generic nature of the individual tools and of the UNIX shell environment, automated experiments can be prepared and executed rapidly in response to unanticipated needs or new ideas. Examples are given of application of this methodology to beam motion characterization, beam-position-monitor offset measurements, and klystron characterization

  17. Practical Unix and Internet Security

    CERN Document Server

    Garfinkel, Simson; Spafford, Gene

    2003-01-01

    When Practical Unix Security was first published more than a decade ago, it became an instant classic. Crammed with information about host security, it saved many a Unix system administrator from disaster. The second edition added much-needed Internet security coverage and doubled the size of the original volume. The third edition is a comprehensive update of this very popular book - a companion for the Unix/Linux system administrator who needs to secure his or her organization's system, networks, and web presence in an increasingly hostile world. Focusing on the four most popular Unix varia

  18. Switching the JLab Accelerator Operations Environment from an HP-UX Unix-based to a PC/Linux-based environment

    International Nuclear Information System (INIS)

    Mcguckin, Theodore

    2008-01-01

    The Jefferson Lab Accelerator Controls Environment (ACE) was predominantly based on the HP-UX Unix platform from 1987 through the summer of 2004. During this period the Accelerator Machine Control Center (MCC) underwent a major renovation which included introducing Redhat Enterprise Linux machines, first as specialized process servers and then gradually as general login servers. As computer programs and scripts required to run the accelerator were modified, and inherent problems with the HP-UX platform compounded, more development tools became available for use with Linux and the MCC began to be converted over. In May 2008 the last HP-UX Unix login machine was removed from the MCC, leaving only a few Unix-based remote-login servers still available. This presentation will explore the process of converting an operational Control Room environment from the HP-UX to Linux platform as well as the many hurdles that had to be overcome throughout the transition period

  19. Mac OS X for Unix Geeks (Leopard)

    CERN Document Server

    Rothman, Ernest E; Rosen, Rich

    2009-01-01

    If you've been lured to Mac OS X because of its Unix roots, this invaluable book serves as a bridge between Apple's Darwin OS and the more traditional Unix systems. The new edition offers a complete tour of Mac OS X's Unix shell for Leopard and Tiger, and helps you find the facilities that replace or correspond to standard Unix utilities. Learn how to compile code, link to libraries, and port Unix software to Mac OS X and much more with this concise guide.

  20. Mac OS X Tiger for Unix Geeks

    CERN Document Server

    Jepson, Brian

    2005-01-01

    If you're one of the many Unix developers drawn to Mac OS X for its Unix core, you'll find yourself in surprisingly unfamiliar territory. Unix and Mac OS X are kissing cousins, but there are enough pitfalls and minefields in going from one to another that even a Unix guru can stumble, and most guides to Mac OS X are written for Mac aficionados. For a Unix developer, approaching Tiger from the Mac side is a bit like learning Russian by reading the Russian side of a Russian-English dictionary. Fortunately, O'Reilly has been the Unix authority for over 25 years, and in Mac OS X Tiger for Unix Gee

  1. Real-time UNIX in HEP data acquisition

    International Nuclear Information System (INIS)

    Buono, S.; Gaponenko, I.; Jones, R.; Mapelli, L.; Mornacchi, G.; Prigent, D.; Sanchez-Corral, E.; Skiadelli, M.; Toppers, A.; Duval, P.Y.; Ferrato, D.; Le Van Suu, A.; Qian, Z.; Rondot, C.; Ambrosini, G.; Fumagalli, G.; Aguer, M.; Huet, M.

    1994-01-01

    Today's experimentation in high energy physics is characterized by an increasing need for sensitivity to rare phenomena and complex physics signatures, which require the use of huge and sophisticated detectors and consequently a high performance readout and data acquisition. Multi-level triggering, hierarchical data collection and an always increasing amount of processing power, distributed throughout the data acquisition layers, will impose a number of features on the software environment, especially the need for a high level of standardization. Real-time UNIX seems, today, the best solution for the platform independence, operating system interface standards and real-time features necessary for data acquisition in HEP experiments. We present the results of the evaluation, in a realistic application environment, of a Real-Time UNIX operating system: the EP/LX real-time UNIX system. ((orig.))

  2. The UNIX/XENIX Advantage: Applications in Libraries.

    Science.gov (United States)

    Gordon, Kelly L.

    1988-01-01

    Discusses the application of the UNIX/XENIX operating system to support administrative office automation functions--word processing, spreadsheets, database management systems, electronic mail, and communications--at the Central Michigan University Libraries. Advantages and disadvantages of the XENIX operating system and system configuration are…

  3. Multimedia Synchronization and UNIX-or-If Multimedia Support is the Problem, Is UNIX the Solution?

    NARCIS (Netherlands)

    D.C.A. Bulterman (Dick); G. van Rossum (Guido); D.T. Winter (Dik)

    1991-01-01

    htmlabstractThis paper considers the role of UNIX in supporting multimedia applications. In particular, we consider the ability of the UNIX operating system (in general) and the UNIX I/O system (in particular) to support the synchronization of a number of high-bandwidth data sets that must be

  4. From a UNIX to a PC Based SCADA System

    CERN Document Server

    Momal, F

    1999-01-01

    In order to facilitate the development of supervisory applications involved in slow process control (such as cryogenic control), the LHC/IAS Group (Equipment Controls Group) opted, a few years ago, for an industrial SCADA package which runs on UNIXÒ platforms. However, to reduce costs and following market trends, it has been decided to move over to a PC-based package. Several processes relating to the testing of the prototypes of the LHC magnets are already controlled in this way. However, it was still necessary to provide all the services previously available to the users, for example, data archiving in central databases, real-time access through the Web, automatic GSM calls, etc. This paper presents the advantages and drawbacks of a PC-based package versus a Unix-based system. It also lists the criteria used in the market survey to arrive at the final selection, as well as, the overall architecture, highlighting the developments needed to integrate the package into the global computing environment.

  5. Protection against hostile algorithms in UNIX software

    Science.gov (United States)

    Radatti, Peter V.

    1996-03-01

    Protection against hostile algorithms contained in Unix software is a growing concern without easy answers. Traditional methods used against similar attacks in other operating system environments such as MS-DOS or Macintosh are insufficient in the more complex environment provided by Unix. Additionally, Unix provides a special and significant problem in this regard due to its open and heterogeneous nature. These problems are expected to become both more common and pronounced as 32 bit multiprocess network operating systems become popular. Therefore, the problems experienced today are a good indicator of the problems and the solutions that will be experienced in the future, no matter which operating system becomes predominate.

  6. Learning Unix for Mac OS X Tiger Unlock the Power of Unix

    CERN Document Server

    Taylor, Dave

    2005-01-01

    Thoroughly revised and updated for Mac OS X Tiger, this new edition introduces Mac users to the Terminal application and shows you how to navigate the command interface, explore hundreds of Unix applications that come with the Mac, and, most importantly, how to take advantage of both the Mac and Unix interfaces. If you want to master the command-line, this gentle guide to using Unix on Mac OS X Tiger is well worth its cover price

  7. A secure file manager for UNIX

    Energy Technology Data Exchange (ETDEWEB)

    DeVries, R.G.

    1990-12-31

    The development of a secure file management system for a UNIX-based computer facility with supercomputers and workstations is described. Specifically, UNIX in its usual form does not address: (1) Operation which would satisfy rigorous security requirements. (2) Online space management in an environment where total data demands would be many times the actual online capacity. (3) Making the file management system part of a computer network in which users of any computer in the local network could retrieve data generated on any other computer in the network. The characteristics of UNIX can be exploited to develop a portable, secure file manager which would operate on computer systems ranging from workstations to supercomputers. Implementation considerations making unusual use of UNIX features, rather than requiring extensive internal system changes, are described, and implementation using the Cray Research Inc. UNICOS operating system is outlined.

  8. A UNIX-based prototype biomedical virtual image processor

    International Nuclear Information System (INIS)

    Fahy, J.B.; Kim, Y.

    1987-01-01

    The authors have developed a multiprocess virtual image processor for the IBM PC/AT, in order to maximize image processing software portability for biomedical applications. An interprocess communication scheme, based on two-way metacode exchange, has been developed and verified for this purpose. Application programs call a device-independent image processing library, which transfers commands over a shared data bridge to one or more Autonomous Virtual Image Processors (AVIP). Each AVIP runs as a separate process in the UNIX operating system, and implements the device-independent functions on the image processor to which it corresponds. Application programs can control multiple image processors at a time, change the image processor configuration used at any time, and are completely portable among image processors for which an AVIP has been implemented. Run-time speeds have been found to be acceptable for higher level functions, although rather slow for lower level functions, owing to the overhead associated with sending commands and data over the shared data bridge

  9. A small Unix-based data acquisition system

    International Nuclear Information System (INIS)

    Engberg, D.; Glanzman, T.

    1993-06-01

    The proposed SLAC B Factory detector plans to use Unix-based machines for all aspects of computing, including real-time data acquisition and experimental control. An R ampersand D program has been established to investigate the use of Unix in the various aspects of experimental computation. Earlier R ampersand D work investigated the basic real-time aspects of the IBM RS/6000 workstation running AIX. The next step in this R ampersand D is the construction of prototype data acquisition system which attempts to exercise many of the features needed in the final on-line system in a realistic situation. For this project, we have combined efforts with a team studying the use of novel cell designs and gas mixtures in a new prototype drift chamber

  10. gLExec: gluing grid computing to the Unix world

    Science.gov (United States)

    Groep, D.; Koeroo, O.; Venekamp, G.

    2008-07-01

    The majority of compute resources in todays scientific grids are based on Unix and Unix-like operating systems. In this world, user and user-group management are based around the concepts of a numeric 'user ID' and 'group ID' that are local to the resource. In contrast, grid concepts of user and group management are centered around globally assigned identifiers and VO membership, structures that are independent of any specific resource. At the fabric boundary, these 'grid identities' have to be translated to Unix user IDs. New job submission methodologies, such as job-execution web services, community-deployed local schedulers, and the late binding of user jobs in a grid-wide overlay network of 'pilot jobs', push this fabric boundary ever further down into the resource. gLExec, a light-weight (and thereby auditable) credential mapping and authorization system, addresses these issues. It can be run both on fabric boundary, as part of an execution web service, and on the worker node in a late-binding scenario. In this contribution we describe the rationale for gLExec, how it interacts with the site authorization and credential mapping frameworks such as LCAS, LCMAPS and GUMS, and how it can be used to improve site control and traceability in a pilot-job system.

  11. gLExec: gluing grid computing to the Unix world

    International Nuclear Information System (INIS)

    Groep, D; Koeroo, O; Venekamp, G

    2008-01-01

    The majority of compute resources in todays scientific grids are based on Unix and Unix-like operating systems. In this world, user and user-group management are based around the concepts of a numeric 'user ID' and 'group ID' that are local to the resource. In contrast, grid concepts of user and group management are centered around globally assigned identifiers and VO membership, structures that are independent of any specific resource. At the fabric boundary, these 'grid identities' have to be translated to Unix user IDs. New job submission methodologies, such as job-execution web services, community-deployed local schedulers, and the late binding of user jobs in a grid-wide overlay network of 'pilot jobs', push this fabric boundary ever further down into the resource. gLExec, a light-weight (and thereby auditable) credential mapping and authorization system, addresses these issues. It can be run both on fabric boundary, as part of an execution web service, and on the worker node in a late-binding scenario. In this contribution we describe the rationale for gLExec, how it interacts with the site authorization and credential mapping frameworks such as LCAS, LCMAPS and GUMS, and how it can be used to improve site control and traceability in a pilot-job system

  12. UNIX-based operating systems robustness evaluation

    Science.gov (United States)

    Chang, Yu-Ming

    1996-01-01

    Robust operating systems are required for reliable computing. Techniques for robustness evaluation of operating systems not only enhance the understanding of the reliability of computer systems, but also provide valuable feed- back to system designers. This thesis presents results from robustness evaluation experiments on five UNIX-based operating systems, which include Digital Equipment's OSF/l, Hewlett Packard's HP-UX, Sun Microsystems' Solaris and SunOS, and Silicon Graphics' IRIX. Three sets of experiments were performed. The methodology for evaluation tested (1) the exception handling mechanism, (2) system resource management, and (3) system capacity under high workload stress. An exception generator was used to evaluate the exception handling mechanism of the operating systems. Results included exit status of the exception generator and the system state. Resource management techniques used by individual operating systems were tested using programs designed to usurp system resources such as physical memory and process slots. Finally, the workload stress testing evaluated the effect of the workload on system performance by running a synthetic workload and recording the response time of local and remote user requests. Moderate to severe performance degradations were observed on the systems under stress.

  13. FEAT - FAILURE ENVIRONMENT ANALYSIS TOOL (UNIX VERSION)

    Science.gov (United States)

    Pack, G.

    1994-01-01

    saved as a library file which represents a generic digraph structure for a class of components. The Generate Model feature can then use library files to generate digraphs for every component listed in the modeling tables, and these individual digraph files can be used in a variety of ways to speed generation of complete digraph models. FEAT contains a preprocessor which performs transitive closure on the digraph. This multi-step algorithm builds a series of phantom bridges, or gates, that allow accurate bi-directional processing of digraphs. This preprocessing can be time-consuming, but once preprocessing is complete, queries can be answered and displayed within seconds. A UNIX X-Windows port of version 3.5 of FEAT, XFEAT, is also available to speed the processing of digraph models created on the Macintosh. FEAT v3.6, which is only available for the Macintosh, has some report generation capabilities which are not available in XFEAT. For very large integrated systems, FEAT can be a real cost saver in terms of design evaluation, training, and knowledge capture. The capability of loading multiple digraphs and schematics into FEAT allows modelers to build smaller, more focused digraphs. Typically, each digraph file will represent only a portion of a larger failure scenario. FEAT will combine these files and digraphs from other modelers to form a continuous mathematical model of the system's failure logic. Since multiple digraphs can be cumbersome to use, FEAT ties propagation results to schematic drawings produced using MacDraw II (v1.1v2 or later) or MacDraw Pro. This makes it easier to identify single and double point failures that may have to cross several system boundaries and multiple engineering disciplines before creating a hazardous condition. FEAT v3.6 for the Macintosh is written in C-language using Macintosh Programmer's Workshop C v3.2. It requires at least a Mac II series computer running System 7 or System 6.0.8 and 32 Bit QuickDraw. It also requires a math

  14. CERN's common Unix and X terminal environment

    International Nuclear Information System (INIS)

    Cass, Tony

    1996-01-01

    The Desktop Infrastructure Group of CERN's Computing and Networks Division has developed a Common Unix and X Terminal Environment to case the migration to Unix based Interactive Computing. The CUTE architecture relies on a distributed flesystem - currently Transarc's AFS - to enable essentially interchangeable client workstation to access both home directory and program files transparently. Additionally, we provide a suite of programs to configure workstations for CUTE and to ensure continued compatibility. This paper describes the different components and the development of the CUTE architecture. (author)

  15. Python for Unix and Linux system administration

    CERN Document Server

    Gift, Noah

    2007-01-01

    Python is an ideal language for solving problems, especially in Linux and Unix networks. With this pragmatic book, administrators can review various tasks that often occur in the management of these systems, and learn how Python can provide a more efficient and less painful way to handle them. Each chapter in Python for Unix and Linux System Administration presents a particular administrative issue, such as concurrency or data backup, and presents Python solutions through hands-on examples. Once you finish this book, you'll be able to develop your own set of command-line utilities with Pytho

  16. The Fermi Unix Environment - Dealing with Adolescence

    Science.gov (United States)

    Pordes, Ruth; Nicholls, Judy; Wicks, Matt

    Fermilab's Computing Division started early in the definition implemention and promulgation of a common environment for Users across the Laboratory's UNIX platforms and installations. Based on our experience over nearly five years, we discuss the status of the effort ongoing developments and needs, some analysis of where we could have done better, and identify future directions to allow us to provide better and more complete service to our customers. In particular, with the power of the new PCs making enthusiastic converts of physicists to the pc world, we are faced with the challenge of expanding the paradigm to non-UNIX platforms in a uniform and consistent way.

  17. Script-viruses Attacks on UNIX OS

    Directory of Open Access Journals (Sweden)

    D. M. Mikhaylov

    2010-06-01

    Full Text Available In this article attacks on UNIX OS are considered. Currently antivirus developers are concentrated on protecting systems from viruses that are most common and attack popular operating systems. If the system or its components are not often attacked then the antivirus products are not protecting these components as it is not profitable. The same situation is with script-viruses for UNIX OS as most experts consider that it is impossible for such viruses to get enough rights to attack. Nevertheless the main conclusion of this article is the fact that such viruses can be very powerful and can attack systems and get enough rights.

  18. UNIX at high energy physics Laboratories

    Energy Technology Data Exchange (ETDEWEB)

    Silverman, Alan

    1994-03-15

    With more and more high energy physics Laboratories ''downsizing'' from large central proprietary mainframe computers towards distributed networks, usually involving UNIX operating systems, the need was expressed at the 1991 Computers in HEP (CHEP) Conference to create a group to consider the implications of this trend and perhaps work towards some common solutions to ease the transition for HEP users worldwide.

  19. TkPl_SU: An Open-source Perl Script Builder for Seismic Unix

    Science.gov (United States)

    Lorenzo, J. M.

    2017-12-01

    TkPl_SU (beta) is a graphical user interface (GUI) to select parameters for Seismic Unix (SU) modules. Seismic Unix (Stockwell, 1999) is a widely distributed free software package for processing seismic reflection and signal processing. Perl/Tk is a mature, well-documented and free object-oriented graphical user interface for Perl. In a classroom environment, shell scripting of SU modules engages students and helps focus on the theoretical limitations and strengths of signal processing. However, complex interactive processing stages, e.g., selection of optimal stacking velocities, killing bad data traces, or spectral analysis requires advanced flows beyond the scope of introductory classes. In a research setting, special functionality from other free seismic processing software such as SioSeis (UCSD-NSF) can be incorporated readily via an object-oriented style to programming. An object oriented approach is a first step toward efficient extensible programming of multi-step processes, and a simple GUI simplifies parameter selection and decision making. Currently, in TkPl_SU, Perl 5 packages wrap 19 of the most common SU modules that are used in teaching undergraduate and first-year graduate student classes (e.g., filtering, display, velocity analysis and stacking). Perl packages (classes) can advantageously add new functionality around each module and clarify parameter names for easier usage. For example, through the use of methods, packages can isolate the user from repetitive control structures, as well as replace the names of abbreviated parameters with self-describing names. Moose, an extension of the Perl 5 object system, greatly facilitates an object-oriented style. Perl wrappers are self-documenting via Perl programming document markup language.

  20. UNIX at high energy physics Laboratories

    International Nuclear Information System (INIS)

    Silverman, Alan

    1994-01-01

    With more and more high energy physics Laboratories ''downsizing'' from large central proprietary mainframe computers towards distributed networks, usually involving UNIX operating systems, the need was expressed at the 1991 Computers in HEP (CHEP) Conference to create a group to consider the implications of this trend and perhaps work towards some common solutions to ease the transition for HEP users worldwide

  1. Process control program development

    International Nuclear Information System (INIS)

    Dameron, H.J.

    1985-01-01

    This paper details the development and implementation of a ''Process Control Program'' at Duke Power's three nuclear stations - Oconee, McGuire, and Catawba. Each station is required by Technical Specification to have a ''Process Control Program'' (PCP) to control all dewatering and/or solidification activities for radioactive wastes

  2. Robot welding process control

    Science.gov (United States)

    Romine, Peter L.

    1991-01-01

    This final report documents the development and installation of software and hardware for Robotic Welding Process Control. Primary emphasis is on serial communications between the CYRO 750 robotic welder, Heurikon minicomputer running Hunter & Ready VRTX, and an IBM PC/AT, for offline programming and control and closed-loop welding control. The requirements for completion of the implementation of the Rocketdyne weld tracking control are discussed. The procedure for downloading programs from the Intergraph, over the network, is discussed. Conclusions are made on the results of this task, and recommendations are made for efficient implementation of communications, weld process control development, and advanced process control procedures using the Heurikon.

  3. SPAM- SPECTRAL ANALYSIS MANAGER (UNIX VERSION)

    Science.gov (United States)

    Solomon, J. E.

    1994-01-01

    The Spectral Analysis Manager (SPAM) was developed to allow easy qualitative analysis of multi-dimensional imaging spectrometer data. Imaging spectrometers provide sufficient spectral sampling to define unique spectral signatures on a per pixel basis. Thus direct material identification becomes possible for geologic studies. SPAM provides a variety of capabilities for carrying out interactive analysis of the massive and complex datasets associated with multispectral remote sensing observations. In addition to normal image processing functions, SPAM provides multiple levels of on-line help, a flexible command interpretation, graceful error recovery, and a program structure which can be implemented in a variety of environments. SPAM was designed to be visually oriented and user friendly with the liberal employment of graphics for rapid and efficient exploratory analysis of imaging spectrometry data. SPAM provides functions to enable arithmetic manipulations of the data, such as normalization, linear mixing, band ratio discrimination, and low-pass filtering. SPAM can be used to examine the spectra of an individual pixel or the average spectra over a number of pixels. SPAM also supports image segmentation, fast spectral signature matching, spectral library usage, mixture analysis, and feature extraction. High speed spectral signature matching is performed by using a binary spectral encoding algorithm to separate and identify mineral components present in the scene. The same binary encoding allows automatic spectral clustering. Spectral data may be entered from a digitizing tablet, stored in a user library, compared to the master library containing mineral standards, and then displayed as a timesequence spectral movie. The output plots, histograms, and stretched histograms produced by SPAM can be sent to a lineprinter, stored as separate RGB disk files, or sent to a Quick Color Recorder. SPAM is written in C for interactive execution and is available for two different

  4. Model Process Control Language

    Data.gov (United States)

    National Aeronautics and Space Administration — The MPC (Model Process Control) language enables the capture, communication and preservation of a simulation instance, with sufficient detail that it can be...

  5. Integrating UNIX workstation into existing online data acquisition systems for Fermilab experiments

    International Nuclear Information System (INIS)

    Oleynik, G.

    1991-03-01

    With the availability of cost effective computing prior from multiple vendors of UNIX workstations, experiments at Fermilab are adding such computers to their VMS based online data acquisition systems. In anticipation of this trend, we have extended the software products available in our widely used VAXONLINE and PANDA data acquisition software systems, to provide support for integrating these workstations into existing distributed online systems. The software packages we are providing pave the way for the smooth migration of applications from the current Data Acquisition Host and Monitoring computers running the VMS operating systems, to UNIX based computers of various flavors. We report on software for Online Event Distribution from VAXONLINE and PANDA, integration of Message Reporting Facilities, and a framework under UNIX for experiments to monitor and view the raw event data produced at any level in their DA system. We have developed software that allows host UNIX computers to communicate with intelligent front-end embedded read-out controllers and processor boards running the pSOS operating system. Both RS-232 and Ethernet control paths are supported. This enables calibration and hardware monitoring applications to be migrated to these platforms. 6 refs., 5 figs

  6. Process control device

    International Nuclear Information System (INIS)

    Hayashi, Toshifumi; Kobayashi, Hiroshi.

    1994-01-01

    A process control device comprises a memory device for memorizing a plant operation target, a plant state or a state of equipments related with each other as control data, a read-only memory device for storing programs, a plant instrumentation control device or other process control devices, an input/output device for performing input/output with an operator, and a processing device which conducts processing in accordance with the program and sends a control demand or a display demand to the input/output device. The program reads out control data relative to a predetermined operation target, compares and verify them with actual values to read out control data to be a practice premise condition which is further to be a practice premise condition if necessary, thereby automatically controlling the plant or requiring or displaying input. Practice presuming conditions for the operation target can be examined succesively in accordance with the program without constituting complicated logical figures and AND/OR graphs. (N.H.)

  7. Use of UNIX in large online processor farms

    Science.gov (United States)

    Biel, Joseph R.

    1990-08-01

    There has been a recent rapid increase in the power of RISC computers running the UNIX operating system. Fermilab has begun to make use of these computers in the next generation of offline computer farms. It is also planning to use such computers in online computer farms. Issues involved in constructing online UNIX farms are discussed.

  8. The Vivitron process control

    International Nuclear Information System (INIS)

    Lutz, J.R.; Marsaudon, J.C.

    1989-10-01

    The operation of the VIVITRON electrostatic accelerator designed since 1981 and under construction at the CRN since 1985 needs a dedicated process control set up. The study and design of this control system started in 1987. The electrostatic accelerators are rarely operated by a modern control system. So little knowledge is available in this field. The timing problems are generally weak but the Vivitron specific structure, with seven porticos in the tank and sophisticated beam handling in the terminal, imposes control equipment inside the tank under extreme severe conditions. Several steps are necessary to achieve the full size control system. Some tests in the MP used as a pilot machine supplied practical information about surrounding accelerator conditions inside the tank. They also provided better knowledge of the beam behavior, especially inside the accelerator tube

  9. Multivariate Statistical Process Control

    DEFF Research Database (Denmark)

    Kulahci, Murat

    2013-01-01

    As sensor and computer technology continues to improve, it becomes a normal occurrence that we confront with high dimensional data sets. As in many areas of industrial statistics, this brings forth various challenges in statistical process control (SPC) and monitoring for which the aim...... is to identify “out-of-control” state of a process using control charts in order to reduce the excessive variation caused by so-called assignable causes. In practice, the most common method of monitoring multivariate data is through a statistic akin to the Hotelling’s T2. For high dimensional data with excessive...... amount of cross correlation, practitioners are often recommended to use latent structures methods such as Principal Component Analysis to summarize the data in only a few linear combinations of the original variables that capture most of the variation in the data. Applications of these control charts...

  10. Evaluation of Unix-Based Integrated Office Automation Products.

    Science.gov (United States)

    1994-04-01

    recipient preferences of networked UNIX users. An e-mail directory contains the preferred applications (e.g., FrameMaker , Excel) for each user, and e...Future Not Available (B) FrameMaker (UNIX) E-Optional I/E-Standard Future* (B) Interleaf (UNIX) I/E-Optional I/E-Standard Future* (B) IslandWrite Not...Optional Future Not Available DXF I-Optional Future Not Available (B) EPSI I-Standard Not Available Future (B) FrameMaker (MIF) E-Optional I/E-Standard Not

  11. Work with Apple's Rhapsody Operating System which Allows Simultaneous UNIX Program Development, UNIX Program Execution, and PC Application Execution

    OpenAIRE

    Summers, Don; Riley, Chris; Cremaldi, Lucien; Sanders, David

    2001-01-01

    Over the past decade, UNIX workstations have provided a very powerful program development environment. However, workstations are more expensive than PCs and Macintoshes and require a system manager for day-to-day tasks such as disk backup, adding users, and setting up print queues. Native commercial software for system maintenance and "PC applications" has been lacking under UNIX. Apple's new Rhapsody operating system puts the current MacOS on a NeXT UNIX foundation and adds an enhanced NeXTS...

  12. Development of control and data processing system for CO{sub 2} laser interferometer

    Energy Technology Data Exchange (ETDEWEB)

    Chiba, Shinichi; Kawano, Yasunori; Tsuchiya, Katsuhiko; Inoue, Akira [Japan Atomic Energy Research Inst., Naka, Ibaraki (Japan). Naka Fusion Research Establishment

    2001-11-01

    CO{sub 2} laser interferometer diagnostic has been operating to measure the central electron density in JT-60U plasmas. We have developed a control and data processing system for the CO{sub 2} laser interferometer with flexible functions of data acquisition, data processing and data transfer in accordance with the sequence of JT-60U discharges. This system is mainly composed of two UNIX workstations and CAMAC clusters, in which the high reliability was obtained by sharing the data process functions to the each workstations. Consequently, the control and data processing system becomes to be able to provide electron density data immediately after a JT-60U discharge, routinely. The realtime feedback control of electron density in JT-60U also becomes to be available by using a reference density signal from the CO{sub 2} laser interferometer. (author)

  13. Development of control and data processing system for CO2 laser interferometer

    International Nuclear Information System (INIS)

    Chiba, Shinichi; Kawano, Yasunori; Tsuchiya, Katsuhiko; Inoue, Akira

    2001-11-01

    CO 2 laser interferometer diagnostic has been operating to measure the central electron density in JT-60U plasmas. We have developed a control and data processing system for the CO 2 laser interferometer with flexible functions of data acquisition, data processing and data transfer in accordance with the sequence of JT-60U discharges. This system is mainly composed of two UNIX workstations and CAMAC clusters, in which the high reliability was obtained by sharing the data process functions to the each workstations. Consequently, the control and data processing system becomes to be able to provide electron density data immediately after a JT-60U discharge, routinely. The realtime feedback control of electron density in JT-60U also becomes to be available by using a reference density signal from the CO 2 laser interferometer. (author)

  14. A UNIX device driver for a Translink II Transputer board

    International Nuclear Information System (INIS)

    Wiley, J.C.

    1991-01-01

    A UNIX device driver for a TransLink II Transputer board is described. A complete listing of the code is presented. The device driver allows a transputer array to be used with the A/UX operating system

  15. Introduction of the UNIX International Performance Management Work Group

    Science.gov (United States)

    Newman, Henry

    1993-01-01

    In this paper we presented the planned direction of the UNIX International Performance Management Work Group. This group consists of concerned system developers and users who have organized to synthesize recommendations for standard UNIX performance management subsystem interfaces and architectures. The purpose of these recommendations is to provide a core set of performance management functions and these functions can be used to build tools by hardware system developers, vertical application software developers, and performance application software developers.

  16. Food Processing Control

    Science.gov (United States)

    1997-01-01

    When NASA started plarning for manned space travel in 1959, the myriad challenges of sustaining life in space included a seemingly mundane but vitally important problem: How and what do you feed an astronaut? There were two main concerns: preventing food crumbs from contaminating the spacecraft's atmosphere or floating into sensitive instruments, and ensuring complete freedom from potentially catastrophic disease-producing bacteria, viruses, and toxins. To solve these concerns, NASA enlisted the help of the Pillsbury Company. Pillsbury quickly solved the first problem by coating bite-size foods to prevent crumbling. They developed the hazard analysis and critical control point (HACCP) concept to ensure against bacterial contamination. Hazard analysis is a systematic study of product, its ingredients, processing conditions, handling, storage, packing, distribution, and directions for consumer use to identify sensitive areas that might prove hazardous. Hazard analysis provides a basis for blueprinting the Critical Control Points (CCPs) to be monitored. CCPs are points in the chain from raw materials to the finished product where loss of control could result in unacceptable food safety risks. In early 1970, Pillsbury plants were following HACCP in production of food for Earthbound consumers. Pillsbury's subsequent training courses for Food and Drug Administration (FDA) personnel led to the incorporation of HACCP in the FDA's Low Acid Canned Foods Regulations, set down in the mid-1970s to ensure the safety of all canned food products in the U.S.

  17. Process control by microprocessors

    Energy Technology Data Exchange (ETDEWEB)

    Arndt, W [ed.

    1978-12-01

    Papers from the workshop Process Control by Microprocessors being organized by the Karlsruhe Nuclear Research Center, Project PDV, together with the VDI/VDE-Gesellschaft fuer Mess- und Regelungstechnik are presented. The workshop was held on December 13 and 14, 1978 at the facilities of the Nuclear Research Center. The papers are arranged according to the topics of the workshop; one chapter deals with today's state of the art of microprocessor hardware and software technology; 5 chapters are dedicated to applications. The report also contains papers which will not be presented at the workshop. Both the workshop and the report are expected to improve and distribute the know-how about this modern technology.

  18. AUTOMATED CONTROL AND REAL-TIME DATA PROCESSING OF WIRE SCANNER/HALO SCRAPER MEASUREMENTS

    International Nuclear Information System (INIS)

    Day, L.A.; Gilpatrick, J.D.

    2001-01-01

    The Low-Energy Demonstration Accelerator (LEDA), assembled and operating at Los Alamos National Laboratory, provides the platform for obtaining measurements of high-power proton beam-halo formation. Control system software and hardware have been integrated and customized to enable the production of real-time beam-halo profiles. The Experimental Physics and Industrial Control System (EPICS) hosted on a VXI platform, Interactive Data Language (IDL) programs hosted on UNIX platforms, and LabVIEW (LV) Virtual Instruments hosted on a PC platform have been integrated and customized to provide real-time, synchronous motor control, data acquisition, and data analysis of data acquired through specialized DSP instrumentation. These modules communicate through EPICS Channel Access (CA) communication protocol extensions to control and manage execution flow ensuring synchronous data acquisition and real-time processing of measurement data. This paper describes the software integration and management scheme implemented to produce these real-time beam profiles

  19. Real-time on a standard UNIX workstation?

    International Nuclear Information System (INIS)

    Glanzman, T.

    1992-09-01

    This is a report of an ongoing R ampersand D project which is investigating the use of standard UNIX workstations for the real-time data acquisition from a major new experimental initiative, the SLAC B Factory (PEP II). For this work an IBM RS/6000 workstation running the AIX operating system is used. Real-time extensions to the UNIX operating system are explored and performance measured. These extensions comprise a set of AIX-specific and POSIX-compliant system services. Benchmark comparisons are made with embedded processor technologies. Results are presented for a simple prototype on-line system for laboratory-testing of a new prototype drift chamber

  20. UNIX secure server : a free, secure, and functional server example

    OpenAIRE

    Sastre, Hugo

    2016-01-01

    The purpose of this thesis work was to introduce UNIX server as a personal server but also as a start point for investigation and developing at a professional level. The objective of this thesis was to build a secure server providing not only a FTP server but also an HTTP server and a cloud system for remote backups. OpenBSD was used as the operating system. OpenBSD is a UNIX-like operating system made by hackers for hackers. The difference with other systems that might partially provid...

  1. DWPF process control

    International Nuclear Information System (INIS)

    Heckendoin, F.M. II

    1983-01-01

    The Defense Waste Processing Facility (DWPF) for waste vitrification at the Savannah River Plant (SRP) is in the final design stage. Instrumentation to provide the parameter sensing required to assure the quality of the two-foot-diameter, ten-foot-high waste canister is in the final stage of development. All step of the process and instrumentation are now operating as nearly full-scale prototypes at SRP. Quality will be maintained by assuring that only the intended material enters the canisters, and by sensing the resultant condition of the filled canisters. Primary emphasis will be on instrumentation of the process

  2. Process control in biogas plants

    DEFF Research Database (Denmark)

    Holm-Nielsen, Jens Bo; Oleskowicz-Popiel, Piotr

    2013-01-01

    Efficient monitoring and control of anaerobic digestion (AD) processes are necessary in order to enhance biogas plant performance. The aim of monitoring and controlling the biological processes is to stabilise and optimise the production of biogas. The principles of process analytical technology...

  3. Trusted Unix Working Group (TRUSIX) Rationale for Selecting Access Control List Features for the Unix System

    National Research Council Canada - National Science Library

    1989-01-01

    .... By addressing the class B3 issues, the NCSC believes that this information will also kelp vendors understand how evaluation interpretations will be made at the levels of trust below this class...

  4. Unix version of CALOR89 for calorimeter applications

    International Nuclear Information System (INIS)

    Handler, T.

    1992-01-01

    CALOR89 is a system of coupled Monte Carlo particle transport computer codes which has been successfully employed for the estimation of calorimeter parameters in High Energy Physics. In the past CALOR89 has been running on various IBM machines and on CRAY X-MP at Lawrence Livermore Lab. These machines had non-unix operating systems. In this report we present a UNIX version of CALOR89, which is especially suited for the UNIX work stations. Moreover CALOR89 is also been supplemented with two new program packages which makes it more user friendly. CALPREP is a program for the preparation of the input files for CALOR89 in general geometry and ANALYZ is an analysis package to extract the final results from CALOR89 relevant to calorimeters. This report also provides two script files LCALOR and PCALOR. LCALOR runs CALOR89 sequences of programs and EGS4 for a given configuration sequentially on a single processor and PCALOR concurrently on a multiprocessor unix workstation

  5. Dosimetry and process control for radiation processing

    International Nuclear Information System (INIS)

    Mod Ali, N.

    2002-01-01

    Complete text of publication follows. Accurate radiation dosimetry can provide quality assurance in radiation processing. Considerable relevant experiences in dosimetry by the SSDL-MINT has necessitate the development of methods making measurement at gamma plant traceable to the national standard. It involves the establishment of proper calibration procedure and selection of appropriate transfer system/technique to assure adequate traceability to a primary radiation standard. The effort forms the basis for irradiation process control, the legal approval of the process by the public health authorities (medical product sterilization and food preservation) and the safety and acceptance of the product

  6. Improved operating scenarios of the DIII-D tokamak as a result of the addition of UNIX computer systems

    International Nuclear Information System (INIS)

    Henline, P.A.

    1995-10-01

    The increased use of UNIX based computer systems for machine control, data handling and analysis has greatly enhanced the operating scenarios and operating efficiency of the DRI-D tokamak. This paper will describe some of these UNIX systems and their specific uses. These include the plasma control system, the electron cyclotron heating control system, the analysis of electron temperature and density measurements and the general data acquisition system (which is collecting over 130 Mbytes of data). The speed and total capability of these systems has dramatically affected the ability to operate DIII-D. The improved operating scenarios include better plasma shape control due to the more thorough MHD calculations done between shots and the new ability to see the time dependence of profile data as it relates across different spatial locations in the tokamak. Other analysis which engenders improved operating abilities will be described

  7. Process and apparatus for controlling control rods

    International Nuclear Information System (INIS)

    Gebelin, B.; Couture, R.

    1987-01-01

    This process and apparatus is characterized by 2 methods, for examination of cluster of nuclear control rods. Foucault current analyzer which examines fraction by fraction all the control rods. This examination is made by rotation of the cluster. Doubtful rods are then analysed by ultrasonic probe [fr

  8. Robust control charts in statistical process control

    NARCIS (Netherlands)

    Nazir, H.Z.

    2014-01-01

    The presence of outliers and contaminations in the output of the process highly affects the performance of the design structures of commonly used control charts and hence makes them of less practical use. One of the solutions to deal with this problem is to use control charts which are robust

  9. Microprocessors control of fermentation process

    Energy Technology Data Exchange (ETDEWEB)

    Fawzy, A S; Hinton, O R

    1980-01-01

    This paper presents three schemes for the solution of the optimal control of fermentation process. It also shows the advantages of using microprocessors in controlling and monitoring this process. A linear model of the system is considered. An optimal feedback controller is determined which maintains the states (substrate and organisms concentration) at desired values when the system is subjected to disturbances in the influent substrate and organisms concentration. Simulation results are presented for the three cases.

  10. ADAM (Affordable Desktop Application Manager): a Unix desktop application manager

    International Nuclear Information System (INIS)

    Liebana, M.; Marquina, M.; Ramos, R.

    1996-01-01

    ADAM stands for Affordable Desktop Application Manager. It is a GUI developed at CERN with the aim to ease access to applications. The motivation to develop ADAM came from the unavailability of environments like COSE/CDE and their heavy resource consumption. ADAM has proven to be user friendly: new users are able to customize it to their needs in few minutes. Groups of users may share through ADAM a common application environment. ADAM also integrates the Unix and the PC world. PC users can excess Unix applications in the same way as their usual Windows applications. This paper describes all the ADAM features, how they are used at CERN Public Services, and the future plans for ADAM. (author)

  11. HUMPF [Heterogeneous Unix Montecarlo Production Facility] users guide

    International Nuclear Information System (INIS)

    Cahill, P.; Edgecock, R.; Fisher, S.M.; Gee, C.N.P.; Gordon, J.C.; Kidd, T.; Leake, J.; Rigby, D.J.; Roberts, J.H.C.

    1992-11-01

    The Heterogenous Unix Monte Carlo Production Facility (HUMPF) simplifies the running of particle physics simulation programs on Unix workstations. Monte Carlo is the largest consumer of IBM (CPU) capacity within the Atlas centre at Rutherford Appleton Laboratory (RAL). It is likely that the future computing requirements of the LEP and HERA experiments cannot be satisfied by the IBM 3090 system. HUMPF adds extra capacity, and can be expanded with minimal effort. Monte Carlo programs are CPU-bound, and make little use of the vector or the input/output capacity of the IBM 3090. Such programs are therefore excellent candidates to use the spare capacity of powerful workstations. The main data storage is still handled centrally by the IBM 3090 and its peripherals. The HUMPF facility is suitable for any program with a similar profile. (author)

  12. Statistical Process Control for KSC Processing

    Science.gov (United States)

    Ford, Roger G.; Delgado, Hector; Tilley, Randy

    1996-01-01

    The 1996 Summer Faculty Fellowship Program and Kennedy Space Center (KSC) served as the basis for a research effort into statistical process control for KSC processing. The effort entailed several tasks and goals. The first was to develop a customized statistical process control (SPC) course for the Safety and Mission Assurance Trends Analysis Group. The actual teaching of this course took place over several weeks. In addition, an Internet version of the same course complete with animation and video excerpts from the course when it was taught at KSC was developed. The application of SPC to shuttle processing took up the rest of the summer research project. This effort entailed the evaluation of SPC use at KSC, both present and potential, due to the change in roles for NASA and the Single Flight Operations Contractor (SFOC). Individual consulting on SPC use was accomplished as well as an evaluation of SPC software for KSC use in the future. A final accomplishment of the orientation of the author to NASA changes, terminology, data format, and new NASA task definitions will allow future consultation when the needs arise.

  13. Contribution to data acquisition software of Eurogram and Diamant multi detectors in an Unix/VXWorks environment; Contribution aux logiciels d`acquisition de donnees des multidetecteurs Eurogam et Diamant dans un environnement reparti Unix/VXWorks

    Energy Technology Data Exchange (ETDEWEB)

    Diarra, C

    1994-06-01

    Questions on nuclear matter, need to have new performant equipments. Eurogram is a 4 PI gamma radiations multi detector and a precious tool in gamma spectroscopy, but it is necessary to use a charged particles detector and in this aim Diamant is an Eurogram partner. These two multi detectors needed special software data acquisition systems. The whole of acquisition control and management is based on sun stations with unix system. 56 figs., 64 refs.

  14. UNIX-a solution to the compatibility problem

    International Nuclear Information System (INIS)

    Gulbranson, R.L.

    1983-01-01

    The UNIX operating system (TM Bell Laboratories) has achieved a high degree of popularity in recent years. It is rapidly becoming a defacto standard as the operating system for 16 and 32 bit microcomputers. The adoption of this operating system by the physics community offers several substantial advantages; a portable software environment (editors, file system, etc.), freedom to choose among a variety of higher-level languages for software applications, and computer hardware vendor independence

  15. Montecarlo Simulations for a Lep Experiment with Unix Workstation Clusters

    Science.gov (United States)

    Bonesini, M.; Calegari, A.; Rossi, P.; Rossi, V.

    Modular systems of RISC CPU based computers have been implemented for large productions of Montecarlo simulated events for the DELPHI experiment at CERN. From a pilot system based on DEC 5000 CPU’s, a full size system based on a CONVEX C3820 UNIX supercomputer and a cluster of HP 735 workstations has been put into operation as a joint effort between INFN Milano and CILEA.

  16. Simple Models for Process Control

    Czech Academy of Sciences Publication Activity Database

    Gorez, R.; Klán, Petr

    2011-01-01

    Roč. 22, č. 2 (2011), s. 58-62 ISSN 0929-2268 Institutional research plan: CEZ:AV0Z10300504 Keywords : process model s * PID control * second order dynamics Subject RIV: JB - Sensors, Measurment, Regulation

  17. Quantum Control of Molecular Processes

    CERN Document Server

    Shapiro, Moshe

    2012-01-01

    Written by two of the world's leading researchers in the field, this is a systematic introduction to the fundamental principles of coherent control, and to the underlying physics and chemistry.This fully updated second edition is enhanced by 80% and covers the latest techniques and applications, including nanostructures, attosecond processes, optical control of chirality, and weak and strong field quantum control. Developments and challenges in decoherence-sensitive condensed phase control as well as in bimolecular control are clearly described.Indispensable for atomic, molecular and chemical

  18. Contribution to data acquisition software of Eurogram and Diamant multi detectors in an Unix/VXWorks environment

    International Nuclear Information System (INIS)

    Diarra, C.

    1994-06-01

    Questions on nuclear matter, need to have new performant equipments. Eurogram is a 4 PI gamma radiations multi detector and a precious tool in gamma spectroscopy, but it is necessary to use a charged particles detector and in this aim Diamant is an Eurogram partner. These two multi detectors needed special software data acquisition systems. The whole of acquisition control and management is based on sun stations with unix system. 56 figs., 64 refs

  19. Welding process modelling and control

    Science.gov (United States)

    Romine, Peter L.; Adenwala, Jinen A.

    1993-01-01

    The research and analysis performed, and software developed, and hardware/software recommendations made during 1992 in development of the PC-based data acquisition system for support of Welding Process Modeling and Control is reported. A need was identified by the Metals Processing Branch of NASA Marshall Space Flight Center, for a mobile data aquisition and analysis system, customized for welding measurement and calibration. Several hardware configurations were evaluated and a PC-based system was chosen. The Welding Measurement System (WMS) is a dedicated instrument, strictly for the use of data aquisition and analysis. Although the WMS supports many of the functions associated with the process control, it is not the intention for this system to be used for welding process control.

  20. Monitoring SLAC High Performance UNIX Computing Systems

    International Nuclear Information System (INIS)

    Lettsome, Annette K.

    2005-01-01

    Knowledge of the effectiveness and efficiency of computers is important when working with high performance systems. The monitoring of such systems is advantageous in order to foresee possible misfortunes or system failures. Ganglia is a software system designed for high performance computing systems to retrieve specific monitoring information. An alternative storage facility for Ganglia's collected data is needed since its default storage system, the round-robin database (RRD), struggles with data integrity. The creation of a script-driven MySQL database solves this dilemma. This paper describes the process took in the creation and implementation of the MySQL database for use by Ganglia. Comparisons between data storage by both databases are made using gnuplot and Ganglia's real-time graphical user interface

  1. Processing Controlled PROs in Spanish

    Science.gov (United States)

    Betancort, Moises; Carreiras, Manuel; Acuna-Farina, Carlos

    2006-01-01

    Two experiments were carried out to investigate the processing of the empty category PRO and the time-course of this in Spanish. Eye movements were recorded while participants read sentences in which a matrix clause was followed by a subordinate infinitival clause, so that the subject or the object of the main clause could act as controller of…

  2. STAR- A SIMPLE TOOL FOR AUTOMATED REASONING SUPPORTING HYBRID APPLICATIONS OF ARTIFICIAL INTELLIGENCE (UNIX VERSION)

    Science.gov (United States)

    Borchardt, G. C.

    1994-01-01

    The Simple Tool for Automated Reasoning program (STAR) is an interactive, interpreted programming language for the development and operation of artificial intelligence (AI) application systems. STAR provides an environment for integrating traditional AI symbolic processing with functions and data structures defined in compiled languages such as C, FORTRAN and PASCAL. This type of integration occurs in a number of AI applications including interpretation of numerical sensor data, construction of intelligent user interfaces to existing compiled software packages, and coupling AI techniques with numerical simulation techniques and control systems software. The STAR language was created as part of an AI project for the evaluation of imaging spectrometer data at NASA's Jet Propulsion Laboratory. Programming in STAR is similar to other symbolic processing languages such as LISP and CLIP. STAR includes seven primitive data types and associated operations for the manipulation of these structures. A semantic network is used to organize data in STAR, with capabilities for inheritance of values and generation of side effects. The AI knowledge base of STAR can be a simple repository of records or it can be a highly interdependent association of implicit and explicit components. The symbolic processing environment of STAR may be extended by linking the interpreter with functions defined in conventional compiled languages. These external routines interact with STAR through function calls in either direction, and through the exchange of references to data structures. The hybrid knowledge base may thus be accessed and processed in general by either side of the application. STAR is initially used to link externally compiled routines and data structures. It is then invoked to interpret the STAR rules and symbolic structures. In a typical interactive session, the user enters an expression to be evaluated, STAR parses the input, evaluates the expression, performs any file input

  3. MULTITASKER, Multitasking Kernel for C and FORTRAN Under UNIX

    International Nuclear Information System (INIS)

    Brooks, E.D. III

    1988-01-01

    1 - Description of program or function: MULTITASKER implements a multitasking kernel for the C and FORTRAN programming languages that runs under UNIX. The kernel provides a multitasking environment which serves two purposes. The first is to provide an efficient portable environment for the development, debugging, and execution of production multiprocessor programs. The second is to provide a means of evaluating the performance of a multitasking program on model multiprocessor hardware. The performance evaluation features require no changes in the application program source and are implemented as a set of compile- and run-time options in the kernel. 2 - Method of solution: The FORTRAN interface to the kernel is identical in function to the CRI multitasking package provided for the Cray XMP. This provides a migration path to high speed (but small N) multiprocessors once the application has been coded and debugged. With use of the UNIX m4 macro preprocessor, source compatibility can be achieved between the UNIX code development system and the target Cray multiprocessor. The kernel also provides a means of evaluating a program's performance on model multiprocessors. Execution traces may be obtained which allow the user to determine kernel overhead, memory conflicts between various tasks, and the average concurrency being exploited. The kernel may also be made to switch tasks every cpu instruction with a random execution ordering. This allows the user to look for unprotected critical regions in the program. These features, implemented as a set of compile- and run-time options, cause extra execution overhead which is not present in the standard production version of the kernel

  4. Interfacing ANSYS to user's programs using UNIX shell program

    Energy Technology Data Exchange (ETDEWEB)

    Kim, In Yong; Kim, Beom Shig [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1994-01-01

    It has been considered to be impossible to interface the ANSYS, which is the commercial finite element code and whose program is not open to public, to the other user's program. When the analysis need to be iterated, the user should wait until the analysis is finished and read the ANSYS result to make the input data for every iteration. In this report the direct interfacing techniques between the ANSYS and the other program using UNIX shell programming are proposed. The detail program lists and the application example are also provided. (Author) 19 refs., 6 figs., 7 tabs.

  5. GEMPAK 5.1 - A GENERAL METEOROLOGICAL PACKAGE (UNIX VERSION)

    Science.gov (United States)

    Desjardins, M. L.

    1994-01-01

    GEMPAK is a general meteorological software package developed at NASA/Goddard Space Flight Center. It includes programs to analyze and display surface, upper-air, and gridded data, including model output. There are very general programs to list, edit, and plot data on maps, to display profiles and time series, to draw and fill contours, to draw streamlines, to plot symbols for clouds, sky cover, and pressure tendency, and draw cross sections in the case of gridded data and sounding data. In addition, there are Barnes objective analysis programs to grid surface and upper-air data. The programs include the capabilities to derive meteorological parameters from those found in the dataset, to perform vertical interpolations of sounding data to different coordinate systems, and to compute an extensive set of gridded diagnostic quantities by specifying various nested combinations of scalars and vector arithmetic, algebraic, and differential operators. The GEMPAK 5.1 graphics/transformation subsystem, GEMPLT, provides device-independent graphics. GEMPLT also has the capability to display output in a variety of map projections or overlaid on satellite imagery. GEMPAK 5.1 is written in FORTRAN 77 and C-language and has been implemented on VAX computers under VMS and on computers running the UNIX operating system. During installation and normal use, this package occupies approximately 100Mb of hard disk space. The UNIX version of GEMPAK includes drivers for several graphic output systems including MIT's X Window System (X11,R4), Sun GKS, PostScript (color and monochrome), Silicon Graphics, and others. The VMS version of GEMPAK also includes drivers for several graphic output systems including PostScript (color and monochrome). The VMS version is delivered with the object code for the Transportable Applications Environment (TAE) program, version 4.1 which serves as a user interface. A color monitor is recommended for displaying maps on video display devices. Data for rendering

  6. CLIPS 6.0 - C LANGUAGE INTEGRATED PRODUCTION SYSTEM, VERSION 6.0 (UNIX VERSION)

    Science.gov (United States)

    Donnell, B.

    1994-01-01

    COOL (that is, a rule can pattern match on objects created using COOL). CLIPS 6.0 provides the capability to define functions, overloaded functions, and global variables interactively. In addition, CLIPS can be embedded within procedural code, called as a subroutine, and integrated with languages such as C, FORTRAN and Ada. CLIPS can be easily extended by a user through the use of several well-defined protocols. CLIPS provides several delivery options for programs including the ability to generate stand alone executables or to load programs from text or binary files. CLIPS 6.0 provides support for the modular development and execution of knowledge bases with the defmodule construct. CLIPS modules allow a set of constructs to be grouped together such that explicit control can be maintained over restricting the access of the constructs by other modules. This type of control is similar to global and local scoping used in languages such as C or Ada. By restricting access to deftemplate and defclass constructs, modules can function as blackboards, permitting only certain facts and instances to be seen by other modules. Modules are also used by rules to provide execution control. The CRSV (Cross-Reference, Style, and Verification) utility included with previous version of CLIPS is no longer supported. The capabilities provided by this tool are now available directly within CLIPS 6.0 to aid in the development, debugging, and verification of large rule bases. COSMIC offers four distribution versions of CLIPS 6.0: UNIX (MSC-22433), VMS (MSC-22434), MACINTOSH (MSC-22429), and IBM PC (MSC-22430). Executable files, source code, utilities, documentation, and examples are included on the program media. All distribution versions include identical source code for the command line version of CLIPS 6.0. This source code should compile on any platform with an ANSI C compiler. Each distribution version of CLIPS 6.0, except that for the Macintosh platform, includes an executable for the

  7. Path modeling and process control

    DEFF Research Database (Denmark)

    Høskuldsson, Agnar; Rodionova, O.; Pomerantsev, A.

    2007-01-01

    and having three or more stages. The methods are applied to a process control of a multi-stage production process having 25 variables and one output variable. When moving along the process, variables change their roles. It is shown how the methods of path modeling can be applied to estimate variables...... be performed regarding the foreseeable output property y, and with respect to an admissible range of correcting actions for the parameters of the next stage. In this paper the basic principles of path modeling is presented. The mathematics is presented for processes having only one stage, having two stages...... of the next stage with the purpose of obtaining optimal or almost optimal quality of the output variable. An important aspect of the methods presented is the possibility of extensive graphic analysis of data that can provide the engineer with a detailed view of the multi-variate variation in data....

  8. An integrated approach to process control

    NARCIS (Netherlands)

    Schippers, W.A.J.

    2001-01-01

    The control of production processes is the subject of several disciplines, such as statistical process control (SPC), total productive maintenance (TPM), and automated process control (APC). Although these disciplines are traditionally separated (both in science and in business practice), their

  9. Palantiri: a distributed real-time database system for process control

    International Nuclear Information System (INIS)

    Tummers, B.J.; Heubers, W.P.J.

    1992-01-01

    The medium-energy accelerator MEA, located in Amsterdam, is controlled by a heterogeneous computer network. A large real-time database contains the parameters involved in the control of the accelerator and the experiments. This database system was implemented about ten years ago and has since been extended several times. In response to increased needs the database system has been redesigned. The new database environment, as described in this paper, consists out of two new concepts: (1) A Palantir which is a per machine process that stores the locally declared data and forwards all non local requests for data access to the appropriate machine. It acts as a storage device for data and a looking glass upon the world. (2) Golems: working units that define the data within the Palantir, and that have knowledge of the hardware they control. Applications access the data of a Golem by name (which do resemble Unix path names). The palantir that runs on the same machine as the application handles the distribution of access requests. This paper focuses on the Palantir concept as a distributed data storage and event handling device for process control. (author)

  10. Understanding Unix/Linux programming a guide to theory and practice

    CERN Document Server

    Molay, Bruce

    2003-01-01

    This book explains in a clear and coherent manner how Unix works, how to understand existing Unix programs, and how to design and create new Unix programs. The book is organized by subsystem, each presented in visual terms and explained using vivid metaphors. It breaks the information into manageable parts that can be presented, explained, and mastered. By using case studies and an extremely reader-friendly manner to illustrate complex ideas and concepts, the book covers the basics of systems programming, users, files and manuals, how to read a directory, using 1S, writing PWD, studying STTY, writing a video game, studying SH, environment and shell variables, I/O redirection and pipes, servers and sockets, writing a web server, license servers, and concurrent functions. For Unix system administrators and programmers, network programmers, and others who have used other operating systems and need to learn Unix programming to expand their skill sets.

  11. With Unix under the hood, the Mac has a toehold in the geophysical sector

    Energy Technology Data Exchange (ETDEWEB)

    Roche, P.

    2004-05-01

    A new lease on life in the geophysical sector is predicted for Apple Computer as result of the company's decision first to convert to the Unix System, and then to develop a new operating system called OS X (O-S-Ten) which runs under a version of Unix called Free BSD. While Apple shows no indication of interest to market its hardware and software to the oil industry, at least one oil company, Houston-based Seitel Inc., is using Apple products for its high performance computing and technical desktop applications, such as storing its onshore and offshore 3-D seismic data on Apple's Xserve RAID rack storage system. In another application, Virginia Polytechnic Institute built a supercomputer using 1,100 64-bit G5 Power Macs. The result is the third fastest super-computer in the world, running at a blazing 10.28 teraflops, or 10.28 trillion calculations per second. By all accounts, it is well suited to oilpatch tasks such as seismic data processing and running large-scale simulations such as fluid flow through porous media. Seitel is also interested in Apple's 64-bit G5 computers to run its seismic data processing operations which require large amounts of computing power. While Apple's hardware and software appear to be well adopted to perform oilpatch tasks, proliferation of use of Apple products by oilpatch companies is hindered by the almost complete absence of oilpatch software. Parallel Geoscience Corporation which had long been interested in creating Mac-based software for geoscience applications, is in the process of filling that gap by refocusing its attention on an O-S-Ten version of its Seismic Precessing Workshop{sup T}M{sup (}SPW). How long it will take to make the conversion will depend on customer demand. 2 figs.

  12. From Process Understanding to Process Control

    NARCIS (Netherlands)

    Streefland, M.

    2010-01-01

    A licensed pharmaceutical process is required to be executed within the validated ranges throughout the lifetime of product manufacturing. Changes to the process usually require the manufacturer to demonstrate that the safety and efficacy of the product remains unchanged. Recent changes in the

  13. A real-time data-acquisition and analysis system with distributed UNIX workstations

    International Nuclear Information System (INIS)

    Yamashita, H.; Miyamoto, K.; Maruyama, K.; Hirosawa, H.; Nakayoshi, K.; Emura, T.; Sumi, Y.

    1996-01-01

    A compact data-acquisition system using three RISC/UNIX TM workstations (SUN TM /SPARCstation TM ) with real-time capabilities of monitoring and analysis has been developed for the study of photonuclear reactions with the large-acceptance spectrometer TAGX. One workstation acquires data from memory modules in the front-end electronics (CAMAC and TKO) with a maximum speed of 300 Kbytes/s, where data size times instantaneous rate is 1 Kbyte x 300 Hz. Another workstation, which has real-time capability for run monitoring, gets the data with a buffer manager called NOVA. The third workstation analyzes the data and reconstructs the event. In addition to a general hardware and software description, priority settings and run control by shell scripts are described. This system has recently been used successfully in a two month long experiment. (orig.)

  14. Chemical process control using Mat lab

    International Nuclear Information System (INIS)

    Kang, Sin Chun; Kim, Raeh Yeon; Kim, Yang Su; Oh, Min; Yeo, Yeong Gu; Jung, Yeon Su

    2001-07-01

    This book is about chemical process control, which includes the basis of process control with conception, function, composition of system and summary, change of laplace and linearization, modeling of chemical process, transfer function and block diagram, the first dynamic property of process, the second dynamic property of process, the dynamic property of combined process, control structure of feedback on component of control system, the dynamic property of feedback control loop, stability of closed loop control structure, expression of process, modification and composition of controller, analysis of vibration response and adjustment controller using vibration response.

  15. Opportunities and challenges for process control in process intensification

    NARCIS (Netherlands)

    Nikacevic, N.M.; Huesman, A.E.M.; Hof, Van den P.M.J.; Stankiewicz, A.

    2012-01-01

    This is a review and position article discussing the role and prospective for process control in process intensification. Firstly, the article outlines the classical role of control in process systems, presenting an overview of control systems’ development, from basic PID control to the advanced

  16. The transition of GTDS to the Unix workstation environment

    Science.gov (United States)

    Carter, D.; Metzinger, R.; Proulx, R.; Cefola, P.

    1995-01-01

    Future Flight Dynamics systems should take advantage of the possibilities provided by current and future generations of low-cost, high performance workstation computing environments with Graphical User Interface. The port of the existing mainframe Flight Dynamics systems to the workstation environment offers an economic approach for combining the tremendous engineering heritage that has been encapsulated in these systems with the advantages of the new computing environments. This paper will describe the successful transition of the Draper Laboratory R&D version of GTDS (Goddard Trajectory Determination System) from the IBM Mainframe to the Unix workstation environment. The approach will be a mix of historical timeline notes, descriptions of the technical problems overcome, and descriptions of associated SQA (software quality assurance) issues.

  17. ACHEMA '85: Process control systems

    International Nuclear Information System (INIS)

    Rosskopf, E.

    1985-01-01

    The strategy obviously adopted by the well-established manufacturers is to offer 'easy-to-handle' equipment to gain new customers, and there is a variety of new compact systems or personal computers being put on the market. The changes and improvements within the processing sector proceed more or less in silence; high-capacity storage devices and multiprocessor configurations are obtainable at a moderate price, offering a greater variety of basic functions and enhanced control possibilities. Redundancy problems are handled with greater flexibility, and batch programs are advancing. Data communication has become a common feature, transmission speed and bus length have been improved. Important improvements have been made with regard to data display; even medium-sized equipment now offer the possibility of making dynamic flow-sheets and reserving space for process history display, and the hierarchy of displays has been considerably simplified. The user software also has been made more easy, 'fill-in-the-blancs' is the prevailing motto for dialog configurations, and such big terms as process computer' or 'programming skill' are passing into oblivion. (orig./HP) [de

  18. Controlling the digital transfer process

    Science.gov (United States)

    Brunner, Felix

    1997-02-01

    The accuracy of today's color management systems fails to satisfy the requirements of the graphic arts market. A first explanation for this is that color calibration charts on which these systems rely, because of print technical reasons, are subject to color deviations and inconsistencies. A second reason is that colorimetry describes the human visual perception of color differences and has no direct relation to the rendering technology itself of a proofing or printing device. The author explains that only firm process control of the many parameters in offset printing by means of a system as for example EUROSTANDARD System Brunner, can lead to accurate and consistent calibration of scanner, display, proof and print. The same principles hold for the quality management of digital presses.

  19. Fuzzy control of pressurizer dynamic process

    International Nuclear Information System (INIS)

    Ming Zhedong; Zhao Fuyu

    2006-01-01

    Considering the characteristics of pressurizer dynamic process, the fuzzy control system that takes the advantages of both fuzzy controller and PID controller is designed for the dynamic process in pressurizer. The simulation results illustrate this type of composite control system is with better qualities than those of single fuzzy controller and single PID controller. (authors)

  20. Business Intelligence in Process Control

    Science.gov (United States)

    Kopčeková, Alena; Kopček, Michal; Tanuška, Pavol

    2013-12-01

    The Business Intelligence technology, which represents a strong tool not only for decision making support, but also has a big potential in other fields of application, is discussed in this paper. Necessary fundamental definitions are offered and explained to better understand the basic principles and the role of this technology for company management. Article is logically divided into five main parts. In the first part, there is the definition of the technology and the list of main advantages. In the second part, an overview of the system architecture with the brief description of separate building blocks is presented. Also, the hierarchical nature of the system architecture is shown. The technology life cycle consisting of four steps, which are mutually interconnected into a ring, is described in the third part. In the fourth part, analytical methods incorporated in the online analytical processing and data mining used within the business intelligence as well as the related data mining methodologies are summarised. Also, some typical applications of the above-mentioned particular methods are introduced. In the final part, a proposal of the knowledge discovery system for hierarchical process control is outlined. The focus of this paper is to provide a comprehensive view and to familiarize the reader with the Business Intelligence technology and its utilisation.

  1. Process control using modern systems of information processing

    International Nuclear Information System (INIS)

    Baldeweg, F.

    1984-01-01

    Modern digital automation techniques allow the application of demanding types of process control. These types of process control are characterized by their belonging to higher levels in a multilevel model. Functional and technical aspects of the performance of digital automation plants are presented and explained. A modern automation system is described considering special procedures of process control (e.g. real time diagnosis)

  2. FASTRAN II - FATIGUE CRACK GROWTH STRUCTURAL ANALYSIS (UNIX VERSION)

    Science.gov (United States)

    Newman, J. C.

    1994-01-01

    loads may be either tensile or compressive. Several standardized aircraft flight-load histories, such as TWIST, Mini-TWIST, FALSTAFF, Inverted FALSTAFF, Felix and Gaussian, are included as options. FASTRAN II also includes two other methods that will help the user input spectrum load histories. The two methods are: (1) a list of stress points, and (2) a flight-by-flight history of stress points. Examples are provided in the user manual. Developed as a research program, FASTRAN II has successfully predicted crack growth in many metallic materials under various aircraft spectrum loading. A computer program DKEFF which is a part of the FASTRAN II package was also developed to analyze crack growth rate data from laboratory specimens to obtain the effective stress-intensity factor against crack growth rate relations used in FASTRAN II. FASTRAN II is written in standard FORTRAN 77. It has been successfully compiled and implemented on Sun4 series computers running SunOS and on IBM PC compatibles running MS-DOS using the Lahey F77L FORTRAN compiler. Sample input and output data are included with the FASTRAN II package. The UNIX version requires 660K of RAM for execution. The standard distribution medium for the UNIX version (LAR-14865) is a .25 inch streaming magnetic tape cartridge in UNIX tar format. It is also available on a 3.5 inch diskette in UNIX tar format. The standard distribution medium for the MS-DOS version (LAR-14944) is a 5.25 inch 360K MS-DOS format diskette. The contents of the diskette are compressed using the PKWARE archiving tools. The utility to unarchive the files, PKUNZIP.EXE, is included. The program was developed in 1984 and revised in 1992. Sun4 and SunOS are trademarks of Sun Microsystems, Inc. IBM PC is a trademark of International Business Machines Corp. MS-DOS is a trademark of Microsoft, Inc. F77L is a trademark of the Lahey Computer Systems, Inc. UNIX is a registered trademark of AT&T Bell Laboratories. PKWARE and PKUNZIP are trademarks of PKWare

  3. Quality control of static irradiation processing products

    International Nuclear Information System (INIS)

    Bao Jianzhong; Chen Xiulan; Cao Hong; Zhai Jianqing

    2002-01-01

    Based on the irradiation processing practice of the nuclear technique application laboratory of Yangzhou Institute of Agricultural Science, the quality control of irradiation processing products is discussed

  4. Unix Domain Sockets Applied in Android Malware Should Not Be Ignored

    Directory of Open Access Journals (Sweden)

    Xu Jiang

    2018-03-01

    Full Text Available Increasingly, malicious Android apps use various methods to steal private user data without their knowledge. Detecting the leakage of private data is the focus of mobile information security. An initial investigation found that none of the existing security analysis systems can track the flow of information through Unix domain sockets to detect the leakage of private data through such sockets, which can result in zero-day exploits in the information security field. In this paper, we conduct the first systematic study on Unix domain sockets as applied in Android apps. Then, we identify scenarios in which such apps can leak private data through Unix domain sockets, which the existing dynamic taint analysis systems do not catch. Based on these insights, we propose and implement JDroid, a taint analysis system that can track information flows through Unix domain sockets effectively to detect such privacy leaks.

  5. INVITATION TO PERFORM Y2K TESTING UNDER UNIX

    CERN Multimedia

    CERN Y2K Co-ordinator

    1999-01-01

    IntroductionA special AFS cell Ôy2k.cern.chÕ has been established to allow service managers and users to test y2k compliance.In addition to AFS, the cluster consists of machines representing all the Unix flavours in use at CERN (AIX, DUNIX, HP-UX, IRIX, LINUX, and SOLARIS).More information can be obtained from the page: http://wwwinfo.cern.ch/pdp/bis/y2k/y2kplus.htmlTesting scheduleThe cluster will be set to 25 December 1999 on fixed days and then left running for three weeks. This gives people one week to prepare test programs in 1999 and two weeks to check the consequences of passing into year 2000. These fixed dates are set as follows:— 19 May 1999, date set to 25/12/99 (year 2000 starts on 26 May) — 9 June1999, date set to 25/12/99 (year 2000 starts on 16 June)— 30 June 1999, date set to 25/12/99 (year 2000 starts on 7 July)If more than these three sessions are needed an announcement will be made later. RegistrationThe following Web page should be used for r...

  6. DET/MPS - THE GSFC ENERGY BALANCE PROGRAM, DIRECT ENERGY TRANSFER/MULTIMISSION SPACECRAFT MODULAR POWER SYSTEM (UNIX VERSION)

    Science.gov (United States)

    Jagielski, J. M.

    1994-01-01

    The DET/MPS programs model and simulate the Direct Energy Transfer and Multimission Spacecraft Modular Power System in order to aid both in design and in analysis of orbital energy balance. Typically, the DET power system has the solar array directly to the spacecraft bus, and the central building block of MPS is the Standard Power Regulator Unit. DET/MPS allows a minute-by-minute simulation of the power system's performance as it responds to various orbital parameters, focusing its output on solar array output and battery characteristics. While this package is limited in terms of orbital mechanics, it is sufficient to calculate eclipse and solar array data for circular or non-circular orbits. DET/MPS can be adjusted to run one or sequential orbits up to about one week, simulated time. These programs have been used on a variety of Goddard Space Flight Center spacecraft projects. DET/MPS is written in FORTRAN 77 with some VAX-type extensions. Any FORTRAN 77 compiler that includes VAX extensions should be able to compile and run the program with little or no modifications. The compiler must at least support free-form (or tab-delineated) source format and 'do do-while end-do' control structures. DET/MPS is available for three platforms: GSC-13374, for DEC VAX series computers running VMS, is available in DEC VAX Backup format on a 9-track 1600 BPI tape (standard distribution) or TK50 tape cartridge; GSC-13443, for UNIX-based computers, is available on a .25 inch streaming magnetic tape cartridge in UNIX tar format; and GSC-13444, for Macintosh computers running AU/X with either the NKR FORTRAN or AbSoft MacFORTRAN II compilers, is available on a 3.5 inch 800K Macintosh format diskette. Source code and test data are supplied. The UNIX version of DET requires 90K of main memory for execution. DET/MPS was developed in 1990. A/UX and Macintosh are registered trademarks of Apple Computer, Inc. VMS, DEC VAX and TK50 are trademarks of Digital Equipment Corporation. UNIX is a

  7. DUAL-PROCESS, a highly reliable process control system

    International Nuclear Information System (INIS)

    Buerger, L.; Gossanyi, A.; Parkanyi, T.; Szabo, G.; Vegh, E.

    1983-02-01

    A multiprocessor process control system is described. During its development the reliability was the most important aspect because it is used in the computerized control of a 5 MW research reactor. DUAL-PROCESS is fully compatible with the earlier single processor control system PROCESS-24K. The paper deals in detail with the communication, synchronization, error detection and error recovery problems of the operating system. (author)

  8. Control of Pressure Process Using Infineon Microcontroller

    Directory of Open Access Journals (Sweden)

    A. Siddique

    2007-07-01

    Full Text Available The main objective of this paper is to design a cost effective controller for real time implementation of pressure process using Infineon micro controller (SAB 80C517A. Model Identification is performed and it is found to be First Order Plus Dead Time Process (FOPDT. The performance measure is tabulated for different parameter and it is found that Proportional (P controller is suitable for controlling the process.

  9. Multivariate Statistical Process Control Charts: An Overview

    OpenAIRE

    Bersimis, Sotiris; Psarakis, Stelios; Panaretos, John

    2006-01-01

    In this paper we discuss the basic procedures for the implementation of multivariate statistical process control via control charting. Furthermore, we review multivariate extensions for all kinds of univariate control charts, such as multivariate Shewhart-type control charts, multivariate CUSUM control charts and multivariate EWMA control charts. In addition, we review unique procedures for the construction of multivariate control charts, based on multivariate statistical techniques such as p...

  10. Statistical process control for serially correlated data

    NARCIS (Netherlands)

    Wieringa, Jakob Edo

    1999-01-01

    Statistical Process Control (SPC) aims at quality improvement through reduction of variation. The best known tool of SPC is the control chart. Over the years, the control chart has proved to be a successful practical technique for monitoring process measurements. However, its usefulness in practice

  11. Third Dutch Process Security Control Event

    NARCIS (Netherlands)

    Luiijf, H.A.M.

    2009-01-01

    On June 4th, 2009, the third Dutch Process Control Security Event took place in Amsterdam. The event, organised by the Dutch National Infrastructure against Cybercrime (NICC), attracted both Dutch process control experts and members of the European SCADA and Control Systems Information Exchange

  12. Fuzzy Control in the Process Industry

    DEFF Research Database (Denmark)

    Jantzen, Jan; Verbruggen, Henk; Østergaard, Jens-Jørgen

    1999-01-01

    Control problems in the process industry are dominated by non-linear and time-varying behaviour, many inner loops, and much interaction between the control loops. Fuzzy controllers have in some cases nevertheless mimicked the control actions of a human operator. Simple fuzzy controllers can...... be designed starting from PID controllers, and in more complex cases these can be used in connection with model-based predictive control. For high level control and supervisory control several simple controllers can be combined in a priority hierarchy such as the one developed in the cement industry...

  13. Applicability of statistical process control techniques

    NARCIS (Netherlands)

    Schippers, W.A.J.

    1998-01-01

    This paper concerns the application of Process Control Techniques (PCTs) for the improvement of the technical performance of discrete production processes. Successful applications of these techniques, such as Statistical Process Control Techniques (SPC), can be found in the literature. However, some

  14. Wind erosion processes and control

    Science.gov (United States)

    Wind erosion continues to threaten the sustainability of our nations' soil, air, and water resources. To effectively apply conservation systems to prevent wind driven soil loss, an understanding of the fundamental processes of wind erosion is necessary so that land managers can better recognize the ...

  15. Portal monitoring technology control process

    International Nuclear Information System (INIS)

    York, R.L.

    1998-01-01

    Portal monitors are an important part of the material protection, control, and accounting (MPC and A) programs in Russia and the US. Although portal monitors are only a part of an integrated MPC and A system, they are an effective means of controlling the unauthorized movement of special nuclear material (SNM). Russian technical experts have gained experience in the use of SNM portal monitors from US experts ad this has allowed them to use the monitors more effectively. Several Russian institutes and companies are designing and manufacturing SNM portal monitors in Russia. Interactions between Russian and US experts have resulted in improvements to the instruments. SNM portal monitor technology has been effectively transferred from the US to Russia and should be a permanent part of the Russian MPC and A Program. Progress in the implementation of the monitors and improvements to how they are used are discussed

  16. Statistical process control in wine industry using control cards

    OpenAIRE

    Dimitrieva, Evica; Atanasova-Pacemska, Tatjana; Pacemska, Sanja

    2013-01-01

    This paper is based on the research of the technological process of automatic filling of bottles of wine in winery in Stip, Republic of Macedonia. The statistical process control using statistical control card is created. The results and recommendations for improving the process are discussed.

  17. Physical Processes Controlling Earth's Climate

    Science.gov (United States)

    Genio, Anthony Del

    2013-01-01

    As background for consideration of the climates of the other terrestrial planets in our solar system and the potential habitability of rocky exoplanets, we discuss the basic physics that controls the Earths present climate, with particular emphasis on the energy and water cycles. We define several dimensionless parameters relevant to characterizing a planets general circulation, climate and hydrological cycle. We also consider issues associated with the use of past climate variations as indicators of future anthropogenically forced climate change, and recent advances in understanding projections of future climate that might have implications for Earth-like exoplanets.

  18. Improving Instruction Using Statistical Process Control.

    Science.gov (United States)

    Higgins, Ronald C.; Messer, George H.

    1990-01-01

    Two applications of statistical process control to the process of education are described. Discussed are the use of prompt feedback to teachers and prompt feedback to students. A sample feedback form is provided. (CW)

  19. Low Activity Waste Feed Process Control Strategy

    International Nuclear Information System (INIS)

    STAEHR, T.W.

    2000-01-01

    The primary purpose of this document is to describe the overall process control strategy for monitoring and controlling the functions associated with the Phase 1B high-level waste feed delivery. This document provides the basis for process monitoring and control functions and requirements needed throughput the double-shell tank system during Phase 1 high-level waste feed delivery. This document is intended to be used by (1) the developers of the future Process Control Plan and (2) the developers of the monitoring and control system

  20. Natural resources and control processes

    CERN Document Server

    Wang, Mu-Hao; Hung, Yung-Tse; Shammas, Nazih

    2016-01-01

    This edited book has been designed to serve as a natural resources engineering reference book as well as a supplemental textbook. This volume is part of the Handbook of Environmental Engineering series, an incredible collection of methodologies that study the effects of pollution and waste in their three basic forms: gas, solid, and liquid. It complements two other books in the series including Environmental and Natural Resources Engineering and Integrated Natural Resources Management that serve as a basis for advanced study or specialized investigation of the theory and analysis of various natural resources systems. This book covers the management of many waste sources including those from agricultural livestock, deep-wells, industries manufacturing dyes, and municipal solid waste incinerators. The purpose of this book is to thoroughly prepare the reader for understanding the sources, treatment and control methods of toxic wastes shown to have harmful effects on the environment. Chapters provide information ...

  1. Quality Control in Production Processes

    Directory of Open Access Journals (Sweden)

    Prístavka Miroslav

    2016-09-01

    Full Text Available The tools for quality management are used for quality improvement throughout the whole Europe and developed countries. Simple statistics are considered one of the most basic methods. The goal was to apply the simple statistical methods to practice and to solve problems by using them. Selected methods are used for processing the list of internal discrepancies within the organization, and for identification of the root cause of the problem and its appropriate solution. Seven basic quality tools are simple graphical tools, but very effective in solving problems related to quality. They are called essential because they are suitable for people with at least basic knowledge in statistics; therefore, they can be used to solve the vast majority of problems.

  2. Analytical control in metallurgical processes

    International Nuclear Information System (INIS)

    Coedo, A.G.; Dorado, M.T.; Padilla, I.

    1998-01-01

    This paper illustrates the role of analysis in enabling metallurgical industry to meet quality demands. For example, for the steel industry the demands by the automotive, aerospace, power generation, tinplate packaging industries and issue of environment near steel plants. Although chemical analysis technology continues to advance, achieving improved speed, precision and accuracy at lower levels of detection, the competitiveness of manufacturing industry continues to drive property demands at least at the same rate. Narrower specification ranges, lower levels of residual elements and economic pressures prescribe faster process routes, all of which lead to increased demands on the analytical function. These damands are illustrated by examples from several market sectors in which customer issues are considered together with ther analytical implications. (Author) 5 refs

  3. Fundamentals of semiconductor manufacturing and process control

    CERN Document Server

    May, Gary S

    2006-01-01

    A practical guide to semiconductor manufacturing from process control to yield modeling and experimental design Fundamentals of Semiconductor Manufacturing and Process Control covers all issues involved in manufacturing microelectronic devices and circuits, including fabrication sequences, process control, experimental design, process modeling, yield modeling, and CIM/CAM systems. Readers are introduced to both the theory and practice of all basic manufacturing concepts. Following an overview of manufacturing and technology, the text explores process monitoring methods, including those that focus on product wafers and those that focus on the equipment used to produce wafers. Next, the text sets forth some fundamentals of statistics and yield modeling, which set the foundation for a detailed discussion of how statistical process control is used to analyze quality and improve yields. The discussion of statistical experimental design offers readers a powerful approach for systematically varying controllable p...

  4. The Newcastle connection: A software subsystem for constructing distributed UNIX systems

    International Nuclear Information System (INIS)

    Randell, B.

    1985-01-01

    The Newcastle connection is a software subsystem that can be added to each of a set of physically interconnected UNIX or UNIX look-alike systems, so as to construct a distributed system which is functionally indistinguishable at both the user and the program level from a conventional single-processor UNIX system. The techniques used are applicable to a variety and multiplicity of both local and wide area networks, and enable all issues of inter-processor communication, network protocols, etc., to be hidden. A brief account is given of experience with such distributed systems, the first of which was constructed in 1982 using a set of PDP11s running UNIX Version 7, and connected by a Cambridge Ring - since this date the Connection has been used to construct distributed systems based on various other computers and versions of UNIX, both at Newcastle and elsewhere. The final sections compare our scheme to various precursor schemes and discuss its potential relevance to other operating systems. (orig.)

  5. A taxonomy of control in intensified processes

    International Nuclear Information System (INIS)

    Barzin, R.; Abd Shukor, S.R.; Ahmad, A.L.

    2006-01-01

    Process Intensification (PI) is a revolutionary approach to design, development and implementation of process and plant. PI technology offers improved environment in a chemical process in terms of better products, and processes which are safer, cleaner, smaller - and cheaper. PI is a strategy of making dramatic reductions in the size of unit operations within chemical plants, in order to achieve given production objectives. However, PI technology would be handicapped if such system is not properly controlled. There are some foreseeable problems in order to control such processes for instance, dynamic interaction between components that make up a control loop, response time of the instrumentations, availability of proper sensor and etc. In some cases, in order to control these systems, advanced control solutions have been applied i.e. model predictive controllers (MPC) and its different algorithms such as quadratic generalized predictive control (QGPC) and self tuning quadratic generalized predictive control (STQGPC). Nevertheless in some cases simpler solutions could be applied to control such system for example proportional integral controller in the control of reactive distillation systems. As mentioned, conventional control systems like proportional-integral, proportional-integral-derivative (PID) controllers and their different structures can be used in PI systems but due to inherent nonlinearity and fast responsiveness of PI systems, digital controllers-regarding to their robustness-are mostly applied in order to control PI systems. Regarding to the fact that choosing the appropriate control strategy is the most essential part of making PI systems possible to be handle easily, taxonomy of the usage of various control structure in controlling PI systems is proposed. This paper offers an overview and discussion on identifying potential problems of instrumentation in PI technology and available control strategies

  6. Geochemical processes controlling minewater pollution

    International Nuclear Information System (INIS)

    Banks, D.

    2004-01-01

    Minewater is a subset of groundwater, subject to broadly similar hydrochemical processes. In 'normal' groundwaters, access to oxidizing species is poor and acid-base reactions tend to dominate over oxidation reactions. Acid-base reactions such as carbonate dissolution and silicate hydrolysis consume protons and carbon dioxide, and release alkalinity and base cations. In mines, the atmospheric environment is rapidly introduced to the deep reducing geosphere (or vice versa in the case of mine waste deposits). This carries the possibility of intense and rapid oxidation of sulphide minerals such as pyrite, to such an extent that these acid-generating redox reactions may dominate over acid-base 'neutralization' reactions and result in the phenomenon of 'acid rock drainage' (ARD). In ARD, a negative correlation is typically observed between pH and concentrations of many metals and metalloids, base cations and sulphate. This correlation is due to genetic co-variation - generation of protons, sulphate and metals in sulphide weathering reactions, pH-dependent solubility of many ARD-related metals and low pH intensifying carbonate dissolution and silicate hydrolysis to release aluminium, silica and base cations. This paper examines the reactions involved in ARD generation and neutralization, and attempts to clarify key concepts such as pH, Eh, alkalinity, acidity and equilibrium constants. Refs. 42 (author)

  7. Nonparametric predictive inference in statistical process control

    NARCIS (Netherlands)

    Arts, G.R.J.; Coolen, F.P.A.; Laan, van der P.

    2000-01-01

    New methods for statistical process control are presented, where the inferences have a nonparametric predictive nature. We consider several problems in process control in terms of uncertainties about future observable random quantities, and we develop inferences for these random quantities hased on

  8. Nonparametric predictive inference in statistical process control

    NARCIS (Netherlands)

    Arts, G.R.J.; Coolen, F.P.A.; Laan, van der P.

    2004-01-01

    Statistical process control (SPC) is used to decide when to stop a process as confidence in the quality of the next item(s) is low. Information to specify a parametric model is not always available, and as SPC is of a predictive nature, we present a control chart developed using nonparametric

  9. Expert systems in process control systems

    International Nuclear Information System (INIS)

    Wittig, T.

    1987-01-01

    To illustrate where the fundamental difference between expert systems in classical diagnosis and in industrial control lie, the work of process control instrumentation is used as an example for the job of expert systems. Starting from the general process of problem-solving, two classes of expert systems can be defined accordingly. (orig.) [de

  10. PROCESS VARIABILITY REDUCTION THROUGH STATISTICAL PROCESS CONTROL FOR QUALITY IMPROVEMENT

    Directory of Open Access Journals (Sweden)

    B.P. Mahesh

    2010-09-01

    Full Text Available Quality has become one of the most important customer decision factors in the selection among the competing product and services. Consequently, understanding and improving quality is a key factor leading to business success, growth and an enhanced competitive position. Hence quality improvement program should be an integral part of the overall business strategy. According to TQM, the effective way to improve the Quality of the product or service is to improve the process used to build the product. Hence, TQM focuses on process, rather than results as the results are driven by the processes. Many techniques are available for quality improvement. Statistical Process Control (SPC is one such TQM technique which is widely accepted for analyzing quality problems and improving the performance of the production process. This article illustrates the step by step procedure adopted at a soap manufacturing company to improve the Quality by reducing process variability using Statistical Process Control.

  11. Advanced coking process control at Rautaruukki Steel

    Energy Technology Data Exchange (ETDEWEB)

    Ritamaki, O.; Luhtaniemi, H. [Rautaruukki Engineering (Finland)

    1999-12-01

    The paper presents the latest development of the Coking Process Management System (CPMS) at Raahe Steel. The latest third generation system is based on the previous system with the addition of fuzzy logic controllers. (The previous second generation system was based simultaneous feed forward and feedback control.) The system development has resulted in balanced coke oven battery heating, decreased variation in process regulation between shifts and increase of process information for operators. The economic results are very satisfactory. 7 figs.

  12. Engineering Process Monitoring for Control Room Operation

    OpenAIRE

    Bätz, M

    2001-01-01

    A major challenge in process operation is to reduce costs and increase system efficiency whereas the complexity of automated process engineering, control and monitoring systems increases continuously. To cope with this challenge the design, implementation and operation of process monitoring systems for control room operation have to be treated as an ensemble. This is only possible if the engineering of the monitoring information is focused on the production objective and is lead in close coll...

  13. Integrated control system for electron beam processes

    Science.gov (United States)

    Koleva, L.; Koleva, E.; Batchkova, I.; Mladenov, G.

    2018-03-01

    The ISO/IEC 62264 standard is widely used for integration of the business systems of a manufacturer with the corresponding manufacturing control systems based on hierarchical equipment models, functional data and manufacturing operations activity models. In order to achieve the integration of control systems, formal object communication models must be developed, together with manufacturing operations activity models, which coordinate the integration between different levels of control. In this article, the development of integrated control system for electron beam welding process is presented as part of a fully integrated control system of an electron beam plant, including also other additional processes: surface modification, electron beam evaporation, selective melting and electron beam diagnostics.

  14. Control measurement system in purex process

    International Nuclear Information System (INIS)

    Mani, V.V.S.

    1985-01-01

    The dependence of a bulk facility handling Purex Process on the control measurement system for evaluating the process performance needs hardly be emphasized. process control, Plant control, inventory control and quality control are the four components of the control measurement system. The scope and requirements of each component are different and the measurement methods are selected accordingly. However, each measurement system has six important elements. These are described in detail. The quality assurance programme carried out by the laboratory as a mechanism through which the quality of measurements is regularly tested and stated in quantitative terms is also explained in terms of internal and external quality assurance, with examples. Suggestions for making the control measurement system more responsive to the operational needs in future are also briefly discussed. (author)

  15. Memory-type control charts in statistical process control

    NARCIS (Netherlands)

    Abbas, N.

    2012-01-01

    Control chart is the most important statistical tool to manage the business processes. It is a graph of measurements on a quality characteristic of the process on the vertical axis plotted against time on the horizontal axis. The graph is completed with control limits that cause variation mark. Once

  16. Optimization and control of metal forming processes

    NARCIS (Netherlands)

    Havinga, Gosse Tjipke

    2016-01-01

    Inevitable variations in process and material properties limit the accuracy of metal forming processes. Robust optimization methods or control systems can be used to improve the production accuracy. Robust optimization methods are used to design production processes with low sensitivity to the

  17. Implementation of the ALEPH detector simulation code using UNIX with on-line graphics display

    International Nuclear Information System (INIS)

    Corden, M.J.; Georgiopoulos, C.H.; Mermikides, M.E.; Streets, J.

    1989-01-01

    GALEPH, the detector simulation program of the ALEPH detector was ported to an ETA10 running under ATandT UNIX System 5. The program on the ETA10 can be driven using standard UNIX socket connections between the ETA and a Silicon Graphics Iris-3020 workstation. The simulated data on the ETA are transferred, using the machine independent binary format EPIO, and displayed on the workstation using a locally developed software package for the visualization of the ALEPH detector. The client (Iris-3020) can also pass parameters to the server (ETA10) and thus interactively change the type of events produced using the same socket connection. (orig.)

  18. UNIX trademark in high energy physics: What we can learn from the initial experiences at Fermilab

    International Nuclear Information System (INIS)

    Butler, J.N.

    1991-03-01

    The reasons why Fermilab decided to support the UNIX operating system are reviewed and placed in the content of an overall model for high energy physics data analysis. The strengths and deficiencies of the UNIX environment for high energy physics are discussed. Fermilab's early experience in dealing with a an ''open'' multivendor environment, both for computers and for peripherals, is described. The human resources required to fully exploit the opportunities are clearly growing. The possibility of keeping the development and support efforts within reasonable bounds may depend on our ability to collaborate or at least to share information even more effectively than we have in the past. 7 refs., 4 figs., 5 tabs

  19. Discrete Control Processes, Dynamic Games and Multicriterion Control Problems

    Directory of Open Access Journals (Sweden)

    Dumitru Lozovanu

    2002-07-01

    Full Text Available The discrete control processes with state evaluation in time of dynamical system is considered. A general model of control problems with integral-time cost criterion by a trajectory is studied and a general scheme for solving such classes of problems is proposed. In addition the game-theoretical and multicriterion models for control problems are formulated and studied.

  20. Testing a Constrained MPC Controller in a Process Control Laboratory

    Science.gov (United States)

    Ricardez-Sandoval, Luis A.; Blankespoor, Wesley; Budman, Hector M.

    2010-01-01

    This paper describes an experiment performed by the fourth year chemical engineering students in the process control laboratory at the University of Waterloo. The objective of this experiment is to test the capabilities of a constrained Model Predictive Controller (MPC) to control the operation of a Double Pipe Heat Exchanger (DPHE) in real time.…

  1. Linearizing control of continuous anaerobic fermentation processes

    Energy Technology Data Exchange (ETDEWEB)

    Babary, J.P. [Centre National d`Etudes Spatiales (CNES), 31 - Toulouse (France). Laboratoire d`Analyse et d`Architecture des Systemes; Simeonov, I. [Institute of Microbiology, Bulgarian Academy of Sciences (Bulgaria); Ljubenova, V. [Institute of Control and System Research, BAS (Country unknown/Code not available); Dochain, D. [Universite Catholique de Louvain (UCL), Louvain-la-Neuve (Belgium)

    1997-09-01

    Biotechnological processes (BTP) involve living organisms. In the anaerobic fermentation (biogas production process) the organic matter is mineralized by microorganisms into biogas (methane and carbon dioxide) in the absence of oxygen. The biogas is an additional energy source. Generally this process is carried out as a continuous BTP. It has been widely used in life process and has been confirmed as a promising method of solving some energy and ecological problems in the agriculture and industry. Because of the very restrictive on-line information the control of this process in continuous mode is often reduced to control of the biogas production rate or the concentration of the polluting organic matter (de-pollution control) at a desired value in the presence of some perturbations. Investigations show that classical linear controllers have good performances only in the linear zone of the strongly non-linear input-output characteristics. More sophisticated robust and with variable structure (VSC) controllers are studied. Due to the strongly non-linear dynamics of the process the performances of the closed loop system may be degrading in this case. The aim of this paper is to investigate different linearizing algorithms for control of a continuous non-linear methane fermentation process using the dilution rate as a control action and taking into account some practical implementation aspects. (authors) 8 refs.

  2. Markov processes and controlled Markov chains

    CERN Document Server

    Filar, Jerzy; Chen, Anyue

    2002-01-01

    The general theory of stochastic processes and the more specialized theory of Markov processes evolved enormously in the second half of the last century. In parallel, the theory of controlled Markov chains (or Markov decision processes) was being pioneered by control engineers and operations researchers. Researchers in Markov processes and controlled Markov chains have been, for a long time, aware of the synergies between these two subject areas. However, this may be the first volume dedicated to highlighting these synergies and, almost certainly, it is the first volume that emphasizes the contributions of the vibrant and growing Chinese school of probability. The chapters that appear in this book reflect both the maturity and the vitality of modern day Markov processes and controlled Markov chains. They also will provide an opportunity to trace the connections that have emerged between the work done by members of the Chinese school of probability and the work done by the European, US, Central and South Ameri...

  3. A NEW BENCHMARK FOR PLANTWIDE PROCESS CONTROL

    Directory of Open Access Journals (Sweden)

    N. Klafke

    Full Text Available Abstract The hydrodealkylation process of toluene (HDA has been used as a case study in a large number of control studies. However, in terms of industrial application, this process has become obsolete and is nowadays superseded by new technologies capable of processing heavy aromatic compounds, which increase the added value of the raw materials, such as the process of transalkylation and disproportionation of toluene (TADP. TADP also presents more complex feed and product streams and challenging operational characteristics both in the reactor and separator sections than in HDA. This work is aimed at proposing the TADP process as a new benchmark for plantwide control studies in lieu of the HAD process. For this purpose, a nonlinear dynamic rigorous model for the TADP process was developed using Aspen Plus™ and Aspen Dynamics™ and industrial conditions. Plantwide control structures (oriented to control and to the process were adapted and applied for the first time for this process. The results show that, even though both strategies are similar in terms of control performance, the optimization of economic factors must still be sought.

  4. Intelligent Controller Design for a Chemical Process

    OpenAIRE

    Mr. Glan Devadhas G; Dr.Pushpakumar S.

    2010-01-01

    Chemical process control is a challenging problem due to the strong on*line non*linearity and extreme sensitivity to disturbances of the process. Ziegler – Nichols tuned PI and PID controllers are found to provide poor performances for higher*order and non–linear systems. This paper presents an application of one*step*ahead fuzzy as well as ANFIS (adaptive*network*based fuzzy inference system) tuning scheme for an Continuous Stirred Tank Reactor CSTR process. The controller is designed based ...

  5. Internal Decoupling in Nonlinear Process Control

    Directory of Open Access Journals (Sweden)

    Jens G. Balchen

    1988-07-01

    Full Text Available A simple method has been investigated for the total or partial removal of the effect of non-linear process phenomena in multi-variable feedback control systems. The method is based upon computing the control variables which will drive the process at desired rates. It is shown that the effect of model errors in the linearization of the process can be partly removed through the use of large feedback gains. In practice there will be limits on how large gains can he used. The sensitivity to parameter errors is less pronounced and the transient behaviour is superior to that of ordinary PI controllers.

  6. Variation and Control of Process Behavior

    International Nuclear Information System (INIS)

    Pawlicki, Todd; Whitaker, Matthew

    2008-01-01

    The purpose of this work was to highlight the importance of controlling process variability for successful quality assurance (QA). We describe the method of statistical process control for characterizing and controlling a process. Traditionally, QA has been performed by comparing some important measurement (e.g., linear accelerator output) against a corresponding specification. Although useful in determining the fitness of a particular measurement, this approach does not provide information about the underlying process behavior over time. A modern view of QA is to consider the time-ordered behavior of a process. Every process displays characteristic behaviors that are independent of the specifications imposed on it. The goal of modern QA is, not only to ensure that a process is on-target, but that it is also operating with minimal variation. This is accomplished by way of a data-driven approach using process behavior charts. The development of process behavior charts, historically known as control charts, and process behavior (action) limits are described. The effect these concepts have on quality management is also discussed

  7. Engineering Process Monitoring for Control Room Operation

    CERN Document Server

    Bätz, M

    2001-01-01

    A major challenge in process operation is to reduce costs and increase system efficiency whereas the complexity of automated process engineering, control and monitoring systems increases continuously. To cope with this challenge the design, implementation and operation of process monitoring systems for control room operation have to be treated as an ensemble. This is only possible if the engineering of the monitoring information is focused on the production objective and is lead in close collaboration of control room teams, exploitation personnel and process specialists. In this paper some principles for the engineering of monitoring information for control room operation are developed at the example of the exploitation of a particle accelerator at the European Laboratory for Nuclear Research (CERN).

  8. Processing implicit control: evidence from reading times

    Directory of Open Access Journals (Sweden)

    Michael eMcCourt

    2015-10-01

    Full Text Available Sentences such as The ship was sunk to collect the insurance exhibit an unusual form of anaphora, implicit control, where neither anaphor nor antecedent is audible. The nonfinite reason clause has an understood subject, PRO, that is anaphoric; here it may be understood as naming the agent of the event of the host clause. Yet since the host is a short passive, this agent is realized by no audible dependent. The putative antecedent to PRO is therefore implicit, which it normally cannot be. What sorts of representations subserve the comprehension of this dependency? Here we present four self-paced reading time studies directed at this question. Previous work showed no processing cost for implicit versus explicit control, and took this to support the view that PRO is linked syntactically to a silent argument in the passive. We challenge this conclusion by reporting that we also find no processing cost for remote implicit control, as in: The ship was sunk. The reason was to collect the insurance. Here the dependency crosses two independent sentences, and so cannot, we argue, be mediated by syntax. Our Experiments 1-4 examined the processing of both implicit (short passive and explicit (active or long passive control in both local and remote configurations. Experiments 3 and 4 added either three days ago or just in order to the local conditions, to control for the distance between the passive and infinitival verbs, and for the predictability of the reason clause, respectively. We replicate the finding that implicit control does not impose an additional processing cost. But critically we show that remote control does not impose a processing cost either. Reading times at the reason clause were never slower when control was remote. In fact they were always faster. Thus efficient processing of local implicit control cannot show that implicit control is mediated by syntax; nor, in turn, that there is a silent but grammatically active argument in passives.

  9. Interface and integration of a silicon graphics UNIX computer with the Encore based SCE SONGS 2/3 simulator

    International Nuclear Information System (INIS)

    Olmos, J.; Lio, P.; Chan, K.S.

    1991-01-01

    The SONGS Unit 2/3 simulator was originally implemented in 1983 on a Master/Slave 32/7780 Encore MPX platform by the Singer-Link Company. In 1986, a 32/9780 MPX Encore computer was incorporated into the simulator computer system to provide the additional CPU processing needed to install the PACE plant monitoring system and to enable the upgrade of the NSSS Simulation to the advanced RETACT/STK models. Since the spring of 1990, the SCE SONGS Nuclear Training Division simulator technical staff, in cooperation with Micro Simulation Inc., has undertaken a project to integrate a Silicon Graphics UNIX based computer with the Encore MPX SONGS 2/3 simulation computer system. In this paper the authors review the objectives, advantages to be gained, software and hardware approaches utilized, and the results so far achieved by the authors' project

  10. Modern control of mineral wool production process

    Directory of Open Access Journals (Sweden)

    Stankov Stanko P.

    2013-01-01

    Full Text Available In this paper, the control of the plant for mineral wool production consisting of a number of the technological units of different sizes and complexity is considered. The application of modern equipment based on PLC (Programmable Logic Controller and SCADA (Supervisory Control And Data Acquisition configuration provides optimal control of technological process. Described supervisory and control system is consisting of a number of units doing decentralized distributed control of technological entities where all possible situation are considered during work of machines and devices, which are installed in electric drive and are protected from technological and electrical accident. Transformer station and diesel engine, raw materials transport and dosage, processes in dome oven, centrifuges, polycondensation (PC chamber, burners, compressor station, binder preparation and dosage, wool cutting, completed panel packing and their transport to storehouse are controlled. Process variables and parameters like as level, flow, velocity, temperature, pressure, etc. are controlled. Control system is doing identification of process states changes, diagnostic and prediction of errors and provides prediction of behavior of control objects when input flows of materials and generates optimal values of control variables due to decreasing downtime and technic - economical requires connected to wool quality to be achieved. Supervisory and control system either eliminates unwanted changes in the production line or restricts them within the allowable limits according to the technology. In this way, the optimization of energy and raw materials consumption and appropriate products quality is achieved, where requirements are satisfied in accordance with process safety and environmental standards. SCADA provides a visual representation of controlled and uncontrolled parts of the technological process, processing alarms and events, monitoring of the changes of relevant

  11. Control of Neutralization Process Using Soft Computing

    Directory of Open Access Journals (Sweden)

    G. Balasubramanian

    2008-03-01

    Full Text Available A novel model-based nonlinear control strategy is proposed using an experimental pH neutralization process. The control strategy involves a non linear neural network (NN model, in the context of internal model control (IMC. When integrated into the internal model control scheme, the resulting controller is shown to have favorable practical implications as well as superior performance. The designed model based online IMC controller was implemented to a laboratory scaled pH process in real time using dSPACE 1104 interface card. The responses of pH and acid flow rate shows good tracking for both the set point and load chances over the entire nonlinear region.

  12. Fuzzy systems for process identification and control

    International Nuclear Information System (INIS)

    Gorrini, V.; Bersini, H.

    1994-01-01

    Various issues related to the automatic construction and on-line adaptation of fuzzy controllers are addressed. A Direct Adaptive Fuzzy Control (this is an adaptive control methodology requiring a minimal knowledge of the processes to be coupled with) derived in a way reminiscent of neurocontrol methods, is presented. A classical fuzzy controller and a fuzzy realization of a PID controller is discussed. These systems implement a highly non-linear control law, and provide to be quite robust, even in the case of noisy inputs. In order to identify dynamic processes of order superior to one, we introduce a more complex architecture, called Recurrent Fuzzy System, that use some fuzzy internal variables to perform an inferential chaining.I

  13. Improving Accuracy of Processing Through Active Control

    Directory of Open Access Journals (Sweden)

    N. N. Barbashov

    2016-01-01

    Full Text Available An important task of modern mathematical statistics with its methods based on the theory of probability is a scientific estimate of measurement results. There are certain costs under control, and under ineffective control when a customer has got defective products these costs are significantly higher because of parts recall.When machining the parts, under the influence of errors a range scatter of part dimensions is offset towards the tolerance limit. To improve a processing accuracy and avoid defective products involves reducing components of error in machining, i.e. to improve the accuracy of machine and tool, tool life, rigidity of the system, accuracy of the adjustment. In a given time it is also necessary to adapt machine.To improve an accuracy and a machining rate there, currently  become extensively popular various the in-process gaging devices and controlled machining that uses adaptive control systems for the process monitoring. Improving the accuracy in this case is compensation of a majority of technological errors. The in-cycle measuring sensors (sensors of active control allow processing accuracy improvement by one or two quality and provide a capability for simultaneous operation of several machines.Efficient use of in-cycle measuring sensors requires development of methods to control the accuracy through providing the appropriate adjustments. Methods based on the moving average, appear to be the most promising for accuracy control since they include data on the change in some last measured values of the parameter under control.

  14. Statistical process control for residential treated wood

    Science.gov (United States)

    Patricia K. Lebow; Timothy M. Young; Stan Lebow

    2017-01-01

    This paper is the first stage of a study that attempts to improve the process of manufacturing treated lumber through the use of statistical process control (SPC). Analysis of industrial and auditing agency data sets revealed there are differences between the industry and agency probability density functions (pdf) for normalized retention data. Resampling of batches of...

  15. Integrated Process Design, Control and Analysis of Intensified Chemical Processes

    DEFF Research Database (Denmark)

    Mansouri, Seyed Soheil

    chemical processes; for example, intensified processes such as reactive distillation. Most importantly, it identifies and eliminates potentially promising design alternatives that may have controllability problems later. To date, a number of methodologies have been proposed and applied on various problems......, that the same principles that apply to a binary non-reactive compound system are valid also for a binary-element or a multi-element system. Therefore, it is advantageous to employ the element based method for multicomponent reaction-separation systems. It is shown that the same design-control principles...

  16. Welding process decoupling for improved control

    International Nuclear Information System (INIS)

    Hardt, D.E.; Eagar, T.W.; Lang, J.H.; Jones, L.

    1993-01-01

    The Gas Metal Arc Welding Process is characterized by many important process outputs, all of which should be controlled to ensure consistent high performance joints. However, application of multivariable control methods is confounded by the strong physical coupling of typical outputs of bead shape and thermal properties. This coupling arises from the three dimensional thermal diffusion processes inherent in welding, and cannot be overcome without significant process modification. This paper presents data on the extent of coupling of the process, and proposes process changes to overcome such strong output coupling. Work in rapid torch vibration to change the heat input distribution is detailed, and methods for changing the heat balance between base and fill material heat are described

  17. Supporting Cross-Organizational Process Control

    Science.gov (United States)

    Angelov, Samuil; Vonk, Jochem; Vidyasankar, Krishnamurthy; Grefen, Paul

    E-contracts express the rights and obligations of parties through a formal, digital representation of the contract provisions. In process intensive relationships, e-contracts contain business processes that a party promises to perform for the counter party, optionally allowing monitoring of the execution of the promised processes. In this paper, we describe an approach in which the counter party is allowed to control the process execution. This approach will lead to more flexible and efficient business relations which are essential in the context of modern, highly dynamic and complex collaborations among companies. We present a specification of the process controls available to the consumer and their support in the private process specification of the provider.

  18. Addressing the Digital Divide in Contemporary Biology: Lessons from Teaching UNIX.

    Science.gov (United States)

    Mangul, Serghei; Martin, Lana S; Hoffmann, Alexander; Pellegrini, Matteo; Eskin, Eleazar

    2017-10-01

    Life and medical science researchers increasingly rely on applications that lack a graphical interface. Scientists who are not trained in computer science face an enormous challenge analyzing high-throughput data. We present a training model for use of command-line tools when the learner has little to no prior knowledge of UNIX. Copyright © 2017 Elsevier Ltd. All rights reserved.

  19. Microcrystalline silicon deposition: Process stability and process control

    International Nuclear Information System (INIS)

    Donker, M.N. van den; Kilper, T.; Grunsky, D.; Rech, B.; Houben, L.; Kessels, W.M.M.; Sanden, M.C.M. van de

    2007-01-01

    Applying in situ process diagnostics, we identified several process drifts occurring in the parallel plate plasma deposition of microcrystalline silicon (μc-Si:H). These process drifts are powder formation (visible from diminishing dc-bias and changing spatial emission profile on a time scale of 10 0 s), transient SiH 4 depletion (visible from a decreasing SiH emission intensity on a time scale of 10 2 s), plasma heating (visible from an increasing substrate temperature on a time scale of 10 3 s) and a still puzzling long-term drift (visible from a decreasing SiH emission intensity on a time scale of 10 4 s). The effect of these drifts on the crystalline volume fraction in the deposited films is investigated by selected area electron diffraction and depth-profiled Raman spectroscopy. An example shows how the transient depletion and long-term drift can be prevented by suitable process control. Solar cells deposited using this process control show enhanced performance. Options for process control of plasma heating and powder formation are discussed

  20. Human factors challenges for advanced process control

    International Nuclear Information System (INIS)

    Stubler, W.F.; O'Hara, J..M.

    1996-01-01

    New human-system interface technologies provide opportunities for improving operator and plant performance. However, if these technologies are not properly implemented, they may introduce new challenges to performance and safety. This paper reports the results from a survey of human factors considerations that arise in the implementation of advanced human-system interface technologies in process control and other complex systems. General trends were identified for several areas based on a review of technical literature and a combination of interviews and site visits with process control organizations. Human factors considerations are discussed for two of these areas, automation and controls

  1. Fault Tolerant Control Using Gaussian Processes and Model Predictive Control

    Directory of Open Access Journals (Sweden)

    Yang Xiaoke

    2015-03-01

    Full Text Available Essential ingredients for fault-tolerant control are the ability to represent system behaviour following the occurrence of a fault, and the ability to exploit this representation for deciding control actions. Gaussian processes seem to be very promising candidates for the first of these, and model predictive control has a proven capability for the second. We therefore propose to use the two together to obtain fault-tolerant control functionality. Our proposal is illustrated by several reasonably realistic examples drawn from flight control.

  2. Controlling Laboratory Processes From A Personal Computer

    Science.gov (United States)

    Will, H.; Mackin, M. A.

    1991-01-01

    Computer program provides natural-language process control from IBM PC or compatible computer. Sets up process-control system that either runs without operator or run by workers who have limited programming skills. Includes three smaller programs. Two of them, written in FORTRAN 77, record data and control research processes. Third program, written in Pascal, generates FORTRAN subroutines used by other two programs to identify user commands with device-driving routines written by user. Also includes set of input data allowing user to define user commands to be executed by computer. Requires personal computer operating under MS-DOS with suitable hardware interfaces to all controlled devices. Also requires FORTRAN 77 compiler and device drivers written by user.

  3. Modeling and Advanced Control for Sustainable Process ...

    Science.gov (United States)

    This book chapter introduces a novel process systems engineering framework that integrates process control with sustainability assessment tools for the simultaneous evaluation and optimization of process operations. The implemented control strategy consists of a biologically-inspired, multi-agent-based method. The sustainability and performance assessment of process operating points is carried out using the U.S. E.P.A.’s GREENSCOPE assessment tool that provides scores for the selected economic, material management, environmental and energy indicators. The indicator results supply information on whether the implementation of the controller is moving the process towards a more sustainable operation. The effectiveness of the proposed framework is illustrated through a case study of a continuous bioethanol fermentation process whose dynamics are characterized by steady-state multiplicity and oscillatory behavior. This book chapter contribution demonstrates the application of novel process control strategies for sustainability by increasing material management, energy efficiency, and pollution prevention, as needed for SHC Sustainable Uses of Wastes and Materials Management.

  4. Intelligent Predictive Control of Nonlienar Processes Using

    DEFF Research Database (Denmark)

    Nørgård, Peter Magnus; Sørensen, Paul Haase; Poulsen, Niels Kjølstad

    1996-01-01

    This paper presents a novel approach to design of generalized predictive controllers (GPC) for nonlinear processes. A neural network is used for modelling the process and a gain-scheduling type of GPC is subsequently designed. The combination of neural network models and predictive control has...... frequently been discussed in the neural network community. This paper proposes an approximate scheme, the approximate predictive control (APC), which facilitates the implementation and gives a substantial reduction in the required amount of computations. The method is based on a technique for extracting...... linear models from a nonlinear neural network and using them in designing the control system. The performance of the controller is demonstrated in a simulation study of a pneumatic servo system...

  5. Novel strategies for control of fermentation processes

    DEFF Research Database (Denmark)

    Mears, Lisa

    to highly optimised industrial host strains. The focus of this project is instead on en-gineering of the process. The question to be answered in this thesis is, given a highly optimised industrial host strain, how can we operate the fermentation process in order to maximise the productivity of the system...... (2012). This model describes the fungal processes operated in the fermentation pilot plant at Novozymes A/S. This model is investigated using uncertainty analysis methods in order to as-sess the applicability to control applications. A mechanistic model approach is desirable, as it is a predictive....... This provides a prediction of the future trajectory of the process, so that it is possible to guide the system to the desired target mass. The control strategy is applied on-line at 550L scale in the Novozymes A/S fermentation pilot plant, and the method is challenged with four different sets of process...

  6. Application of artificial intelligence in process control

    CERN Document Server

    Krijgsman, A

    1993-01-01

    This book is the result of a united effort of six European universities to create an overall course on the appplication of artificial intelligence (AI) in process control. The book includes an introduction to key areas including; knowledge representation, expert, logic, fuzzy logic, neural network, and object oriented-based approaches in AI. Part two covers the application to control engineering, part three: Real-Time Issues, part four: CAD Systems and Expert Systems, part five: Intelligent Control and part six: Supervisory Control, Monitoring and Optimization.

  7. Multivariable adaptive control of bio process

    Energy Technology Data Exchange (ETDEWEB)

    Maher, M.; Bahhou, B.; Roux, G. [Centre National de la Recherche Scientifique (CNRS), 31 - Toulouse (France); Maher, M. [Faculte des Sciences, Rabat (Morocco). Lab. de Physique

    1995-12-31

    This paper presents a multivariable adaptive control of a continuous-flow fermentation process for the alcohol production. The linear quadratic control strategy is used for the regulation of substrate and ethanol concentrations in the bioreactor. The control inputs are the dilution rate and the influent substrate concentration. A robust identification algorithm is used for the on-line estimation of linear MIMO model`s parameters. Experimental results of a pilot-plant fermenter application are reported and show the control performances. (authors) 8 refs.

  8. Microeconomics of process control in semiconductor manufacturing

    Science.gov (United States)

    Monahan, Kevin M.

    2003-06-01

    Process window control enables accelerated design-rule shrinks for both logic and memory manufacturers, but simple microeconomic models that directly link the effects of process window control to maximum profitability are rare. In this work, we derive these links using a simplified model for the maximum rate of profit generated by the semiconductor manufacturing process. We show that the ability of process window control to achieve these economic objectives may be limited by variability in the larger manufacturing context, including measurement delays and process variation at the lot, wafer, x-wafer, x-field, and x-chip levels. We conclude that x-wafer and x-field CD control strategies will be critical enablers of density, performance and optimum profitability at the 90 and 65nm technology nodes. These analyses correlate well with actual factory data and often identify millions of dollars in potential incremental revenue and cost savings. As an example, we show that a scatterometry-based CD Process Window Monitor is an economically justified, enabling technology for the 65nm node.

  9. The monitoring and control of TRUEX processes

    International Nuclear Information System (INIS)

    Regalbuto, M.C.; Misra, B.; Chamberlain, D.B.; Leonard, R.A.; Vandegrift, G.F.

    1992-04-01

    The Generic TRUEX Model (GTM) was used to design a flowsheet for the TRUEX solvent extraction process that would be used to determine its instrumentation and control requirements. Sensitivity analyses of the key process variables, namely, the aqueous and organic flow rates, feed compositions, and the number of contactor stages, were carried out to assess their impact on the operation of the TRUEX process. Results of these analyses provide a basis for the selection of an instrument and control system and the eventual implementation of a control algorithm. Volume Two of this report is an evaluation of the instruments available for measuring many of the physical parameters. Equations that model the dynamic behavior of the TRUEX process have been generated. These equations can be used to describe the transient or dynamic behavior of the process for a given flowsheet in accordance with the TRUEX model. Further work will be done with the dynamic model to determine how and how quickly the system responds to various perturbations. The use of perturbation analysis early in the design stage will lead to a robust flowsheet, namely, one that will meet all process goals and allow for wide control bounds. The process time delay, that is, the speed with which the system reaches a new steady state, is an important parameter in monitoring and controlling a process. In the future, instrument selection and point-of-variable measurement, now done using the steady-state results reported here, will be reviewed and modified as necessary based on this dynamic method of analysis

  10. Neural PID Control Strategy for Networked Process Control

    Directory of Open Access Journals (Sweden)

    Jianhua Zhang

    2013-01-01

    Full Text Available A new method with a two-layer hierarchy is presented based on a neural proportional-integral-derivative (PID iterative learning method over the communication network for the closed-loop automatic tuning of a PID controller. It can enhance the performance of the well-known simple PID feedback control loop in the local field when real networked process control applied to systems with uncertain factors, such as external disturbance or randomly delayed measurements. The proposed PID iterative learning method is implemented by backpropagation neural networks whose weights are updated via minimizing tracking error entropy of closed-loop systems. The convergence in the mean square sense is analysed for closed-loop networked control systems. To demonstrate the potential applications of the proposed strategies, a pressure-tank experiment is provided to show the usefulness and effectiveness of the proposed design method in network process control systems.

  11. Improving industrial process control systems security

    CERN Document Server

    Epting, U; CERN. Geneva. TS Department

    2004-01-01

    System providers are today creating process control systems based on remote connectivity using internet technology, effectively exposing these systems to the same threats as corporate computers. It is becoming increasingly difficult and costly to patch/maintain the technical infrastructure monitoring and control systems to remove these vulnerabilities. A strategy including risk assessment, security policy issues, service level agreements between the IT department and the controls engineering groups must be defined. In addition an increased awareness of IT security in the controls system engineering domain is needed. As consequence of these new factors the control system architectures have to take into account security requirements, that often have an impact on both operational aspects as well as on the project and maintenance cost. Manufacturers of industrial control system equipment do however also propose progressively security related solutions that can be used for our active projects. The paper discusses ...

  12. A methodology to describe process control requirements

    International Nuclear Information System (INIS)

    Carcagno, R.; Ganni, V.

    1994-01-01

    This paper presents a methodology to describe process control requirements for helium refrigeration plants. The SSC requires a greater level of automation for its refrigeration plants than is common in the cryogenics industry, and traditional methods (e.g., written descriptions) used to describe process control requirements are not sufficient. The methodology presented in this paper employs tabular and graphic representations in addition to written descriptions. The resulting document constitutes a tool for efficient communication among the different people involved in the design, development, operation, and maintenance of the control system. The methodology is not limited to helium refrigeration plants, and can be applied to any process with similar requirements. The paper includes examples

  13. Process and control systems for composites manufacturing

    Science.gov (United States)

    Tsiang, T. H.; Wanamaker, John L.

    1992-01-01

    A precise control of composite material processing would not only improve part quality, but it would also directly reduce the overall manufacturing cost. The development and incorporation of sensors will help to generate real-time information for material processing relationships and equipment characteristics. In the present work, the thermocouple, pressure transducer, and dielectrometer technologies were investigated. The monitoring sensors were integrated with the computerized control system in three non-autoclave fabrication techniques: hot-press, self contained tool (self heating and pressurizing), and pressure vessel). The sensors were implemented in the parts and tools.

  14. Functional graphical languages for process control

    International Nuclear Information System (INIS)

    1996-01-01

    A wide variety of safety systems are in use today in the process industries. Most of these systems rely on control software using procedural programming languages. This study investigates the use of functional graphical languages for controls in the process industry. Different vendor proprietary software and languages are investigated and evaluation criteria are outlined based on ability to meet regulatory requirements, reference sites involving applications with similar safety concerns, QA/QC procedures, community of users, type and user-friendliness of the man-machine interface, performance of operational code, and degree of flexibility. (author) 16 refs., 4 tabs

  15. Apparatus and process for controlling fluidized beds

    Science.gov (United States)

    Rehmat, Amirali G.; Patel, Jitendra G.

    1985-10-01

    An apparatus and process for control and maintenance of fluidized beds under non-steady state conditions. An ash removal conduit is provided for removing solid particulates from a fluidized bed separate from an ash discharge conduit in the lower portion of the grate supporting such a bed. The apparatus and process of this invention is particularly suitable for use in ash agglomerating fluidized beds and provides control of the fluidized bed before ash agglomeration is initiated and during upset conditions resulting in stable, sinter-free fluidized bed maintenance.

  16. Aspects of parallel processing and control engineering

    OpenAIRE

    McKittrick, Brendan J

    1991-01-01

    The concept of parallel processing is not a new one, but the application of it to control engineering tasks is a relatively recent development, made possible by contemporary hardware and software innovation. It has long been accepted that, if properly orchestrated several processors/CPUs when combined can form a powerful processing entity. What prevented this from being implemented in commercial systems was the adequacy of the microprocessor for most tasks and hence the expense of a multi-pro...

  17. Ventilation equations for improved exothermic process control.

    Science.gov (United States)

    McKernan, John L; Ellenbecker, Michael J

    2007-04-01

    Exothermic or heated processes create potentially unsafe work environments for an estimated 5-10 million American workers each year. Excessive heat and process contaminants have the potential to cause acute health effects such as heat stroke, and chronic effects such as manganism in welders. Although millions of workers are exposed to exothermic processes, insufficient attention has been given to continuously improving engineering technologies for these processes to provide effective and efficient control. Currently there is no specific occupational standard established by OSHA regarding exposure to heat from exothermic processes, therefore it is important to investigate techniques that can mitigate known and potential adverse occupational health effects. The current understanding of engineering controls for exothermic processes is primarily based on a book chapter written by W. C. L. Hemeon in 1955. Improvements in heat transfer and meteorological theory necessary to design improved process controls have occurred since this time. The research presented involved a review of the physical properties, heat transfer and meteorological theories governing buoyant air flow created by exothermic processes. These properties and theories were used to identify parameters and develop equations required for the determination of buoyant volumetric flow to assist in improving ventilation controls. Goals of this research were to develop and describe a new (i.e. proposed) flow equation, and compare it to currently accepted ones by Hemeon and the American Conference of Governmental Industrial Hygienists (ACGIH). Numerical assessments were conducted to compare solutions from the proposed equations for plume area, mean velocity and flow to those from the ACGIH and Hemeon. Parameters were varied for the dependent variables and solutions from the proposed, ACGIH, and Hemeon equations for plume area, mean velocity and flow were analyzed using a randomized complete block statistical

  18. Evaluation of control strategies in forming processes

    Directory of Open Access Journals (Sweden)

    Calmano Stefan

    2015-01-01

    Full Text Available Products of forming processes are subject to quality fluctuations due to uncertainty in semi-finished part properties as well as process conditions and environment. An approach to cope with these uncertainties is the implementation of a closed-loop control taking into account the actual product properties measured by sensors or estimated by a mathematical process model. Both methods of uncertainty control trade off with a financial effort. In case of sensor integration the effort is the cost of the sensor including signal processing as well as the design and manufacturing effort for integration. In case of an estimation model the effort is mainly determined by the time and knowledge needed to derive the model, identify the parameters and implement the model into the PLC. The risk of mismatch between model and reality as well as the risk of wrong parameter identification can be assumed as additional uncertainty (model uncertainty. This paper evaluates controlled and additional uncertainty by taking into account process boundary conditions like the degree of fluctuations in semi-finished part properties. The proposed evaluation is demonstrated by the analysis of exemplary processes.

  19. Training change control process at Cernavoda NPP

    International Nuclear Information System (INIS)

    Valache, Cornelia Mariana

    2005-01-01

    The paper presents the process of 'Training Change Control' at Cernavoda NPP. This process is a systematic approach that allows determination of the most effective training and/or non-training solutions for challenges that may influence the content and conditions for a training program or course. Changes may be the result of: - response to station systems or equipment modifications; - new or revised procedures; - regulatory requirements; - external organizations requirements; - internal evaluations meaning feedback from trainees, trainers, management or post-training evaluations; - self-assessments; - station condition reports; - operating experience (OPEX); - modifications of job scope; - management input. The Training Change Control Process at Cernavoda NPP includes the following aspects. The first step is the identification of all the initiating factors for a potential training change. Then, retain only those, which could have an impact on training and classify them in two categories: as deficiencies or as enhancement suggestions. The process is different for the two categories. The deficiency category supposes the application of the Training Needs Analysis (TNA) process. This is a performance-oriented process, resulting in more competent employees, solving existing and potential performance problems. By using needs analysis to systematically determine what people or courses and programs are expected to do and gathering data to reveal what they are really doing, we can receive a clear picture of the problem and then we can establish corrective action plans to fix it. The process is supported by plant subjects matter and by training specialists. On the other hand, enhancements suggestions are assessed by designated experienced persons and then are implemented in the training process. Regarding these two types of initiating factors for the training change control process, the final result consists of a training improvement, raising the effectiveness, efficiency or

  20. Optimal control of a CSTR process

    Directory of Open Access Journals (Sweden)

    A. Soukkou

    2008-12-01

    Full Text Available Designing an effective criterion and learning algorithm for find the best structure is a major problem in the control design process. In this paper, the fuzzy optimal control methodology is applied to the design of the feedback loops of an Exothermic Continuous Stirred Tank Reactor system. The objective of design process is to find an optimal structure/gains of the Robust and Optimal Takagi Sugeno Fuzzy Controller (ROFLC. The control signal thus obtained will minimize a performance index, which is a function of the tracking/regulating errors, the quantity of the energy of the control signal applied to the system, and the number of fuzzy rules. The genetic learning is proposed for constructing the ROFLC. The chromosome genes are arranged into two parts, the binary-coded part contains the control genes and the real-coded part contains the genes parameters representing the fuzzy knowledge base. The effectiveness of this chromosome formulation enables the fuzzy sets and rules to be optimally reduced. The performances of the ROFLC are compared to these found by the traditional PD controller with Genetic Optimization (PD_GO. Simulations demonstrate that the proposed ROFLC and PD_GO has successfully met the design specifications.

  1. Novel strategies for control of fermentation processes

    DEFF Research Database (Denmark)

    Mears, Lisa; Stocks, Stuart; Sin, Gürkan

    Bioprocesses are inherently sensitive to fluctuations in processing conditions and must be tightly regulated to maintain cellular productivity. Industrial fermentations are often difficult to replicate across production sites or between facilities as the small operating differences in the equipment...... of a fermentation. Industrial fermentation processes are typically operated in fed batch mode, which also poses specific challenges for process monitoring and control. This is due to many reasons including non-linear behaviour, and a relatively poor understanding of the system dynamics. It is therefore challenging...

  2. FORTRAN data files transference from VAX/VMS to ALPHA/UNIX; Traspaso de ficheros FORTRAN de datos de VAX/VMS a ALPHA/UNIX

    Energy Technology Data Exchange (ETDEWEB)

    Sanchez, E.; Milligen, B. Ph van [CIEMAT (Spain)

    1997-09-01

    Several tools have been developed to access the TJ-IU databases, which currently reside in VAX/VMS servers, from the TJ-II Data Acquisition System DEC ALPHA 8400 server. The TJ-I/TJ-IU databases are not homogeneous and contain several types of data files, namely, SADE, CAMAC and FORTRAN unformatted files. The tools presented in this report allow one to transfer CAMAC and those FORTRAN unformatted files defined herein, from a VAX/VMS server, for data manipulation on the ALPHA/Digital UNIX server. (Author)

  3. Process control for sheet-metal stamping process modeling, controller design and shop-floor implementation

    CERN Document Server

    Lim, Yongseob; Ulsoy, A Galip

    2014-01-01

    Process Control for Sheet-Metal Stamping presents a comprehensive and structured approach to the design and implementation of controllers for the sheet metal stamping process. The use of process control for sheet-metal stamping greatly reduces defects in deep-drawn parts and can also yield large material savings from reduced scrap. Sheet-metal forming is a complex process and most often characterized by partial differential equations that are numerically solved using finite-element techniques. In this book, twenty years of academic research are reviewed and the resulting technology transitioned to the industrial environment. The sheet-metal stamping process is modeled in a manner suitable for multiple-input multiple-output control system design, with commercially available sensors and actuators. These models are then used to design adaptive controllers and real-time controller implementation is discussed. Finally, experimental results from actual shopfloor deployment are presented along with ideas for further...

  4. Dosimetry and control of radiation processing

    International Nuclear Information System (INIS)

    1988-01-01

    Eight invited papers on the general theme of 'Dosimetry and Control of Radiation Processing', presented at a one day symposium held at the National Physical Laboratory, are collected together in this document. Seven of the papers are selected and indexed separately. (author)

  5. Applying interactive control to waste processing operations

    International Nuclear Information System (INIS)

    Grasz, E.L.; Merrill, R.D.; Couture, S.A.

    1992-08-01

    At present waste and residue processing includes steps that require human interaction. The risk of exposure to unknown hazardous materials and the potential for radiation contamination motivates the desire to remove operators from these processes. Technologies that facilitate this include glove box robotics, modular systems for remote and automated servicing, and interactive controls that minimize human intervention. LLNL is developing an automated system which is designed to supplant the operator for glove box tasks, thus protecting the operator from the risk of radiation exposure and minimizing operator-associated waste. Although most of the processing can be automated with minimal human interaction, there are some tasks where intelligent intervention is both desirable and necessary to adapt to Enexpected circumstances and events. These activities require that the operator interact with the process using a remote manipulator which provides or reflects a natural feel to the operator. The remote manipulation system which was developed incorporates sensor fusion and interactive control, and provides the operator with an effective means of controlling the robot in a potentially unknown environment. This paper describes recent accomplishments in technology development and integration, and outlines the future goals of Lawrence Livermore National Laboratory for achieving this integrated interactive control capability

  6. First Dutch Process Control Security Event

    NARCIS (Netherlands)

    Luiijf, H.A.M.

    2008-01-01

    On May 21st , 2008, the Dutch National Infrastructure against Cyber Crime (NICC) organised their first Process Control Security Event. Mrs. Annemarie Zielstra, the NICC programme manager, opened the event. She welcomed the over 100 representatives of key industry sectors. “Earlier studies in the

  7. Inhibition: Mental Control Process or Mental Resource?

    Science.gov (United States)

    Im-Bolter, Nancie; Johnson, Janice; Ling, Daphne; Pascual-Leone, Juan

    2015-01-01

    The current study tested 2 models of inhibition in 45 children with language impairment and 45 children with normally developing language; children were aged 7 to 12 years. Of interest was whether a model of inhibition as a mental-control process (i.e., executive function) or as a mental resource would more accurately reflect the relations among…

  8. Fourth Dutch Process Security Control Event

    NARCIS (Netherlands)

    Luiijf, H.A.M.; Zielstra, A.

    2010-01-01

    On December 1st, 2009, the fourth Dutch Process Control Security Event took place in Baarn, The Netherlands. The security event with the title ‘Manage IT!’ was organised by the Dutch National Infrastructure against Cybercrime (NICC). Mid of November, a group of over thirty people participated in the

  9. Applied Behavior Analysis and Statistical Process Control?

    Science.gov (United States)

    Hopkins, B. L.

    1995-01-01

    Incorporating statistical process control (SPC) methods into applied behavior analysis is discussed. It is claimed that SPC methods would likely reduce applied behavior analysts' intimate contacts with problems and would likely yield poor treatment and research decisions. Cases and data presented by Pfadt and Wheeler (1995) are cited as examples.…

  10. Facts about food irradiation: Controlling the process

    International Nuclear Information System (INIS)

    1991-01-01

    This fact sheet briefly reviews the procedures that exist to control the process of food irradiation. It also summarizes the difficulties in identifying irradiated food, which stem from the fact that irradiation does not physically change the food or cause significant chemical changes in foods. 4 refs

  11. Processes mediating expertise in air traffic control

    NARCIS (Netherlands)

    Van Meeuwen, Ludo; Jarodzka, Halszka; Brand-Gruwel, Saskia; Van Merriënboer, Jeroen; De Bock, Jeano; Kirschner, Paul A.

    2010-01-01

    Van Meeuwen, L., Jarodzka, H., Brand-Gruwel, S., Van Merriënboer, J. J. G., De Bock, J. J. P. R., & Kirschner, P. A. (2010, August). Processes mediating expertise in air traffic control. Meeting of the EARLI SIG6/7 Instructional Design and Learning and Instruction with Computers, Ulm, Germany.

  12. Processes mediating expertise in air traffic control

    NARCIS (Netherlands)

    Van Meeuwen, Ludo; Jarodzka, Halszka; Brand-Gruwel, Saskia; Van Merriënboer, Jeroen; De Bock, Jeano; Kirschner, Paul A.

    2011-01-01

    Van Meeuwen, L. W., Jarodzka, H., Brand-Gruwel, S., Van Merriënboer, J. J. G., De Bock, J. J. P. R., & Kirschner, P. A. (2010, September). Processes mediating expertise in air traffic control. Poster presented at the European Association for Aviation Psychology Conference, Budapest.

  13. Statistical process control for alpha spectroscopy

    Energy Technology Data Exchange (ETDEWEB)

    Richardson, W; Majoras, R E [Oxford Instruments, Inc. P.O. Box 2560, Oak Ridge TN 37830 (United States); Joo, I O; Seymour, R S [Accu-Labs Research, Inc. 4663 Table Mountain Drive, Golden CO 80403 (United States)

    1995-10-01

    Statistical process control(SPC) allows for the identification of problems in alpha spectroscopy processes before they occur, unlike standard laboratory Q C which only identifies problems after a process fails. SPC tools that are directly applicable to alpha spectroscopy include individual X-charts and X-bar charts, process capability plots, and scatter plots. Most scientists are familiar with the concepts the and methods employed by SPC. These tools allow analysis of process bias, precision, accuracy and reproducibility as well as process capability. Parameters affecting instrument performance are monitored and analyzed using SPC methods. These instrument parameters can also be compared to sampling, preparation, measurement, and analysis Q C parameters permitting the evaluation of cause effect relationships. Three examples of SPC, as applied to alpha spectroscopy , are presented. The first example investigates background contamination using averaging to show trends quickly. A second example demonstrates how SPC can identify sample processing problems, analyzing both how and why this problem occurred. A third example illustrates how SPC can predict when an alpha spectroscopy process is going to fail. This allows for an orderly and timely shutdown of the process to perform preventative maintenance, avoiding the need to repeat costly sample analyses. 7 figs., 2 tabs.

  14. Statistical process control for alpha spectroscopy

    International Nuclear Information System (INIS)

    Richardson, W.; Majoras, R.E.; Joo, I.O.; Seymour, R.S.

    1995-01-01

    Statistical process control(SPC) allows for the identification of problems in alpha spectroscopy processes before they occur, unlike standard laboratory Q C which only identifies problems after a process fails. SPC tools that are directly applicable to alpha spectroscopy include individual X-charts and X-bar charts, process capability plots, and scatter plots. Most scientists are familiar with the concepts the and methods employed by SPC. These tools allow analysis of process bias, precision, accuracy and reproducibility as well as process capability. Parameters affecting instrument performance are monitored and analyzed using SPC methods. These instrument parameters can also be compared to sampling, preparation, measurement, and analysis Q C parameters permitting the evaluation of cause effect relationships. Three examples of SPC, as applied to alpha spectroscopy , are presented. The first example investigates background contamination using averaging to show trends quickly. A second example demonstrates how SPC can identify sample processing problems, analyzing both how and why this problem occurred. A third example illustrates how SPC can predict when an alpha spectroscopy process is going to fail. This allows for an orderly and timely shutdown of the process to perform preventative maintenance, avoiding the need to repeat costly sample analyses. 7 figs., 2 tabs

  15. An avoidance layer in hierarchical process control

    International Nuclear Information System (INIS)

    De Callatay, A.

    1994-01-01

    A project of layered software architecture is proposed: a safety-critical real-time non-stop simple kernel system includes a layer avoiding threatening actions from operators or programs in other control systems. Complex process-control applications (such as fuzzy systems) are useful for the smooth operation of the system, optimum productivity, efficient diagnostics, and safe management of degraded modes of operation. Defects in these complex process-control applications do not have an impact on safety if their commands have first to be accepted by a safety-critical module. The development, testing, and certification of complex applications computed in the outside layers can be made simpler and less expensive than for those in the kernel. Avoidance systems use rule-base systems having negative fuzzy conditions and actions. Animal and human behaviour cannot be explained without active avoidance

  16. Minicomputer controlled test system for process control and monitoring systems

    International Nuclear Information System (INIS)

    Worster, L.D.

    A minicomputer controlled test system for testing process control and monitoring systems is described. This system, in service for over one year, has demonstrated that computerized control of such testing has a real potential for expanding the scope of the testing, improving accuracy of testing, and significantly reducing the time required to do the testing. The test system is built around a 16-bit minicomputer with 12K of memory. The system programming language is BASIC with the addition of assembly level routines for communication with the peripheral devices. The peripheral devices include a 100 channel scanner, analog-to-digital converter, visual display, and strip printer. (auth)

  17. Statistical process control for electron beam monitoring.

    Science.gov (United States)

    López-Tarjuelo, Juan; Luquero-Llopis, Naika; García-Mollá, Rafael; Quirós-Higueras, Juan David; Bouché-Babiloni, Ana; Juan-Senabre, Xavier Jordi; de Marco-Blancas, Noelia; Ferrer-Albiach, Carlos; Santos-Serra, Agustín

    2015-07-01

    To assess the electron beam monitoring statistical process control (SPC) in linear accelerator (linac) daily quality control. We present a long-term record of our measurements and evaluate which SPC-led conditions are feasible for maintaining control. We retrieved our linac beam calibration, symmetry, and flatness daily records for all electron beam energies from January 2008 to December 2013, and retrospectively studied how SPC could have been applied and which of its features could be used in the future. A set of adjustment interventions designed to maintain these parameters under control was also simulated. All phase I data was under control. The dose plots were characterized by rising trends followed by steep drops caused by our attempts to re-center the linac beam calibration. Where flatness and symmetry trends were detected they were less-well defined. The process capability ratios ranged from 1.6 to 9.3 at a 2% specification level. Simulated interventions ranged from 2% to 34% of the total number of measurement sessions. We also noted that if prospective SPC had been applied it would have met quality control specifications. SPC can be used to assess the inherent variability of our electron beam monitoring system. It can also indicate whether a process is capable of maintaining electron parameters under control with respect to established specifications by using a daily checking device, but this is not practical unless a method to establish direct feedback from the device to the linac can be devised. Copyright © 2015 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  18. Milestones in screen-based process control

    International Nuclear Information System (INIS)

    Guesnier, G.P.

    1995-01-01

    The German approach is based on the utilisation of the conceptual elements of the PRISCA information system developed by Siemens and on operational experience with screen-based process control in a conventional power plant. In the French approach, the screen-based control room for the N4 plants, designed from scratch, has undergone extensive simulator tests for validation before going into realisation. It is now used in the commissioning phase of the first N4 plants. The design of the control room for the European Pressurized Water Reactor will be based on the common experience of Siemens and Electricite de France. Its main elements are several separate operator workstations, a safety control area used as a back-up for postulated failures of the workstations, and a commonly utilisable plant overview for the operators' coordination. (orig./HP) [de

  19. Expect systems and optimisation in process control

    Energy Technology Data Exchange (ETDEWEB)

    Mamdani, A.; Efstathiou, J. (eds.)

    1986-01-01

    This report brings together recent developments both in expert systems and in optimisation, and deals with current applications in industry. Part One is concerned with Artificial Intellegence in planning and scheduling and with rule-based control implementation. The tasks of control maintenance, rescheduling and planning are each discussed in relation to new theoretical developments, techniques available, and sample applications. Part Two covers model based control techniques in which the control decisions are used in a computer model of the process. Fault diagnosis, maintenance and trouble-shooting are just some of the activities covered. Part Three contains case studies of projects currently in progress, giving details of the software available and the likely future trends. One of these, on qualitative plant modelling as a basis for knowledge-based operator aids in nuclear power stations is indexed separately.

  20. Expert systems and optimisation in process control

    International Nuclear Information System (INIS)

    Mamdani, A.; Efstathiou, J.

    1986-01-01

    This report brings together recent developments both in expert systems and in optimisation, and deals with current applications in industry. Part One is concerned with Artificial Intellegence in planning and scheduling and with rule-based control implementation. The tasks of control maintenance, rescheduling and planning are each discussed in relation to new theoretical developments, techniques available, and sample applications. Part Two covers model based control techniques in which the control decisions are used in a computer model of the process. Fault diagnosis, maintenance and trouble-shooting are just some of the activities covered. Part Three contains case studies of projects currently in progress, giving details of the software available and the likely future trends. One of these, on qualitative plant modelling as a basis for knowledge-based operator aids in nuclear power stations is indexed separately. (author)

  1. Processes subject to integrated pollution control. Petroleum processes: oil refining and associated processes

    International Nuclear Information System (INIS)

    1995-01-01

    This document, part of a series offering guidance on pollution control regulations issued by Her Majesty's Inspectorate of Pollution, (HMIP) focuses on petroleum processes such as oil refining and other associated processes. The various industrial processes used, their associated pollution release routes into the environment and techniques for controlling these releases are all discussed. Environmental quality standards are related to national and international agreements on pollution control and abatement. HMIP's work on air, water and land pollution monitoring is also reported. (UK)

  2. Hydrothermal processing of Hanford tank wastes: Process modeling and control

    International Nuclear Information System (INIS)

    Currier, R.P.

    1994-01-01

    In the Los Alamos National Laboratory (LANL) hydrothermal process, waste streams are first pressurized and heated as they pass through a continuous flow tubular reactor vessel. The waste is maintained at reaction temperature of 300--550 C where organic destruction and sludge reformation occur. This report documents LANL activities in process modeling and control undertaken in FY94 to support hydrothermal process development. Key issues discussed include non-ideal flow patterns (e.g. axial dispersion) and their effect on reactor performance, the use and interpretation of inert tracer experiments, and the use of computational fluid mechanics to evaluate novel hydrothermal reactor designs. In addition, the effects of axial dispersion (and simplifications to rate expressions) on the estimated kinetic parameters are explored by non-linear regression to experimental data. Safety-related calculations are reported which estimate the explosion limits of effluent gases and the fate of hydrogen as it passes through the reactor. Development and numerical solution of a generalized one-dimensional mathematical model is also summarized. The difficulties encountered in using commercially available software to correlate the behavior of high temperature, high pressure aqueous electrolyte mixtures are summarized. Finally, details of the control system and experiments conducted to empirically determine the system response are reported

  3. Process control upgrades yield huge operational improvements

    International Nuclear Information System (INIS)

    Fitzgerald, W.V.

    2001-01-01

    Most nuclear plants in North America were designed and built in the late 60 and 70. The regulatory nature of this industry over the years has made design changes at the plant level difficult, if not impossible, to implement. As a result, many plants in this world region have been getting by on technology that is over 40 years behind the times. What this translates into is that the plants have not been able to take advantage of the huge technology gains that have been made in process control during this period. As a result, most of these plants are much less efficient and productive than they could be. One particular area of the plant that is receiving a lot of attention is the feedwater heaters. These systems were put in place to improve efficiency, but most are not operating correctly. This paper will present a case study where one progressive mid-western utility decided that enough was enough and implemented a process control audit of their heater systems. The audit clearly pointed out the existing problems with the current process control system. It resulted in a proposal for the implementation of a state of the art, digital distributed process control system for the heaters along with a complete upgrade of the level controls and field devices that will stabilize heater levels, resulting in significant efficiency gains and lower maintenance bills. Overall the payback period for this investment should be less than 6 months and the plant is now looking for more opportunities that can provide even bigger gains. (author)

  4. Environmental control costs for oil shale processes

    Energy Technology Data Exchange (ETDEWEB)

    None

    1979-10-01

    The studies reported herein are intended to provide more certainty regarding estimates of the costs of controlling environmental residuals from oil shale technologies being readied for commercial application. The need for this study was evident from earlier work conducted by the Office of Environment for the Department of Energy Oil Shale Commercialization Planning, Environmental Readiness Assessment in mid-1978. At that time there was little reliable information on the costs for controlling residuals and for safe handling of wastes from oil shale processes. The uncertainties in estimating costs of complying with yet-to-be-defined environmental standards and regulations for oil shale facilities are a critical element that will affect the decision on proceeding with shale oil production. Until the regulatory requirements are fully clarified and processes and controls are investigated and tested in units of larger size, it will not be possible to provide definitive answers to the cost question. Thus, the objective of this work was to establish ranges of possible control costs per barrel of shale oil produced, reflecting various regulatory, technical, and financing assumptions. Two separate reports make up the bulk of this document. One report, prepared by the Denver Research Institute, is a relatively rigorous engineering treatment of the subject, based on regulatory assumptions and technical judgements as to best available control technologies and practices. The other report examines the incremental cost effect of more conservative technical and financing alternatives. An overview section is included that synthesizes the products of the separate studies and addresses two variations to the assumptions.

  5. UNIX trademark in high energy physics: What we can learn from the initial experiences at Fermilab

    Energy Technology Data Exchange (ETDEWEB)

    Butler, J.N.

    1991-03-01

    The reasons why Fermilab decided to support the UNIX operating system are reviewed and placed in the content of an overall model for high energy physics data analysis. The strengths and deficiencies of the UNIX environment for high energy physics are discussed. Fermilab's early experience in dealing with a an open'' multivendor environment, both for computers and for peripherals, is described. The human resources required to fully exploit the opportunities are clearly growing. The possibility of keeping the development and support efforts within reasonable bounds may depend on our ability to collaborate or at least to share information even more effectively than we have in the past. 7 refs., 4 figs., 5 tabs.

  6. The statistical process control methods - SPC

    Directory of Open Access Journals (Sweden)

    Floreková Ľubica

    1998-03-01

    Full Text Available Methods of statistical evaluation of quality – SPC (item 20 of the documentation system of quality control of ISO norm, series 900 of various processes, products and services belong amongst basic qualitative methods that enable us to analyse and compare data pertaining to various quantitative parameters. Also they enable, based on the latter, to propose suitable interventions with the aim of improving these processes, products and services. Theoretical basis and applicatibily of the principles of the: - diagnostics of a cause and effects, - Paret analysis and Lorentz curve, - number distribution and frequency curves of random variable distribution, - Shewhart regulation charts, are presented in the contribution.

  7. Statistical process control for radiotherapy quality assurance

    International Nuclear Information System (INIS)

    Pawlicki, Todd; Whitaker, Matthew; Boyer, Arthur L.

    2005-01-01

    Every quality assurance process uncovers random and systematic errors. These errors typically consist of many small random errors and a very few number of large errors that dominate the result. Quality assurance practices in radiotherapy do not adequately differentiate between these two sources of error. The ability to separate these types of errors would allow the dominant source(s) of error to be efficiently detected and addressed. In this work, statistical process control is applied to quality assurance in radiotherapy for the purpose of setting action thresholds that differentiate between random and systematic errors. The theoretical development and implementation of process behavior charts are described. We report on a pilot project is which these techniques are applied to daily output and flatness/symmetry quality assurance for a 10 MV photon beam in our department. This clinical case was followed over 52 days. As part of our investigation, we found that action thresholds set using process behavior charts were able to identify systematic changes in our daily quality assurance process. This is in contrast to action thresholds set using the standard deviation, which did not identify the same systematic changes in the process. The process behavior thresholds calculated from a subset of the data detected a 2% change in the process whereas with a standard deviation calculation, no change was detected. Medical physicists must make decisions on quality assurance data as it is acquired. Process behavior charts help decide when to take action and when to acquire more data before making a change in the process

  8. Access to CAMAC from VxWorks and UNIX in DART

    International Nuclear Information System (INIS)

    Streets, J.; Meadows, J.; Moore, C.

    1995-05-01

    As part of the DART Project the authors have developed a package of software for CAMAC access from UNIX and VxWorks platforms, with support for several hardware interfaces. They report on developments for the CES CBD8210 VME to parallel CAMAC, the Hytec VSD2992 VME to serial CAMAC and Jorway 411S SCSI to parallel and serial CAMAC branch drivers, and give a summary of the timings obtained

  9. San Onofre 2/3 simulator: The move from Unix to Windows

    International Nuclear Information System (INIS)

    Paquette, C.; Desouky, C.; Gagnon, V.

    2006-01-01

    CAE has been developing nuclear power plant (NPP) simulators for over 30 years for customers around the world. While numerous operating systems are used today for simulators, many of the existing simulators were developed to run on workstation-type computers using a variant of the Unix operating system. Today, thanks to the advances in the power and capabilities of Personal Computers (PC's), and because most simulators will eventually need to be upgraded, more and more of these RISC processor-based simulators will be converted to PC-based platforms running either the Windows or Linux operating systems. CAE's multi-platform simulation environment runs on the UNIX Linux and Windows operating systems, enabling simulators to be 'open' and highly interoperable systems using industry-standard software components and methods. The result is simulators that are easier to maintain and modify as reference plants evolve. In early January 2003, CAE set out to upgrade Southern California Edison's San Onofre Unit 2/3 UNIX-based simulator with its latest integrated simulation environment. This environment includes CAE's instructor station Isis, the latest ROSE modeling and runtime tool, as well as the deployment of a new reactor kinetics model (COMET) and new nuclear steam supply system (ANTHEM2000). The chosen simulation platform is PC-based and runs the Windows XP operating system. The main features and achievements of the San Onofre 2/3 Simulator's modernization from RISC/Unix to Intel/Windows XP, running CAE's current simulation environment, is the subject of this paper. (author)

  10. Migration of the UNIX Application for eFAST CANDU Nuclear Power Plant Analyzer

    International Nuclear Information System (INIS)

    Suh, Jae Seung; Sohn, Dae Seong; Kim, Sang Jae; Jeun, Gyoo Dong

    2006-01-01

    Since the mid 1980s, corporate data centers have been moving away from mainframes running dedicated operating systems to mini-computers, often using one or other of the myriad flavors of UNIX. At the same time, the users' experience of these systems has, in many cases, stayed the same, involving text-based interaction with dumb terminals or a terminal-emulation session on a Personal Computer. More recently, IT managers have questioned this approach, and have been looking at changes in the UNIX marketplace and the increasing expense of being tied in to single-vendor software and hardware solutions. The growth of Linux as a lightweight version of UNIX has fueled this interest, raising the number of organizations that are considering a migration to alternative platforms. The various implementations of the UNIX operating system have served industry well, as witnessed by the very large base both of installed systems and large-scale applications installed on those systems. However, there are increasing signs of dissatisfaction with expensive, often proprietary solutions and a growing sense that perhaps the concept of 'big iron' has had its day in the same way as it has for most of the mainframes of the type portrayed in 1970s science fiction films. One of the most extraordinary and unexpected successes of the Intel PC architecture is the extent to which this basic framework has been extended to encompass very large server and data center environments. Large-scale hosting companies are now offering enterprise level services to multiple client companies at availability levels of over 99.99 percent on what are simply racks of relatively cheap PCs. Technologies such as clustering, Network Load Balancing, and Component Load Balancing enable the personal computer to take on and match the levels of throughput, availability, and reliability of all but the most expensive 'big iron' solutions and the supercomputers

  11. Integrated Process Design and Control of Reactive Distillation Processes

    DEFF Research Database (Denmark)

    Mansouri, Seyed Soheil; Sales-Cruz, Mauricio; Huusom, Jakob Kjøbsted

    2015-01-01

    on the element concept, which is used to translate a system of compounds into elements. The operation of the reactive distillation column at the highest driving force and other candidate points is analyzed through analytical solution as well as rigorous open-loop and closed-loop simulations. By application...... of this approach, it is shown that designing the reactive distillation process at the maximum driving force results in an optimal design in terms of controllability and operability. It is verified that the reactive distillation design option is less sensitive to the disturbances in the feed at the highest driving...

  12. Process based analysis of manually controlled drilling processes for bone

    Science.gov (United States)

    Teicher, Uwe; Achour, Anas Ben; Nestler, Andreas; Brosius, Alexander; Lauer, Günter

    2018-05-01

    The machining operation drilling is part of the standard repertoire for medical applications. This machining cycle, which is usually a multi-stage process, generates the geometric element for the subsequent integration of implants, which are screwed into the bone in subsequent processes. In addition to the form, shape and position of the generated drill hole, it is also necessary to use a technology that ensures an operation with minimal damage. A surface damaged by excessive mechanical and thermal energy input shows a deterioration in the healing capacity of implants and represents a structure with complications for inflammatory reactions. The resulting loads are influenced by the material properties of the bone, the used technology and the tool properties. An important aspect of the process analysis is the fact that machining of bone is in most of the cases a manual process that depends mainly on the skills of the operator. This includes, among other things, the machining time for the production of a drill hole, since manual drilling is a force-controlled process. Experimental work was carried out on the bone of a porcine mandible in order to investigate the interrelation of the applied load during drilling. It can be shown that the load application can be subdivided according to the working feed direction. The entire drilling process thus consists of several time domains, which can be divided into the geometry-generating feed motion and a retraction movement of the tool. It has been shown that the removal of the tool from the drill hole has a significant influence on the mechanical load input. This fact is proven in detail by a new evaluation methodology. The causes of this characteristic can also be identified, as well as possible ways of reducing the load input.

  13. Employing expert systems for process control

    International Nuclear Information System (INIS)

    Ahrens, W.

    1987-01-01

    The characteristic features of expert systems are explained in detail, and the systems' application in process control engineering. Four points of main interest are there, namely: Applications for diagnostic tasks, for safety analyses, planning, and training and expert training. For the modelling of the technical systems involved in all four task fields mentioned above, an object-centred approach has shown to be the suitable method, as process control techniques are determined by technical objects that in principle are specified by data sheets, schematic representations, flow charts, and plans. The graphical surface allows these data to be taken into account, so that the object can be displayed in the way best suited to the individual purposes. (orig./GL) [de

  14. Radionuclides for process control and inspection

    International Nuclear Information System (INIS)

    Hadden, R.J.B.

    1987-01-01

    Radiation sources have been used in process control for over 40 years. Their use in inspection, implying visual examination, although of much earlier origin in the form of gamma radiography, is also of recent emergence in the form of tomographic methods. This paper firstly reviews the justification for the continued world-wide usage of isotopic methods. It then reviews a selection of innovative process control applications, based on radiation sources, as illustrations of the present state of the art and also describes recent progress in inspection methods including progress in the development of on-line facilities. For all applications involving radiation sources, careful selection of parameters is required to achieve the highest efficiency compatible with an integrity suitable for the intended application. The paper concludes with a brief discussion of the common principles on which the fabrication of sources is based in order to satisfy national and international safety legislation. (author)

  15. Modular and Adaptive Control of Sound Processing

    Science.gov (United States)

    van Nort, Douglas

    This dissertation presents research into the creation of systems for the control of sound synthesis and processing. The focus differs from much of the work related to digital musical instrument design, which has rightly concentrated on the physicality of the instrument and interface: sensor design, choice of controller, feedback to performer and so on. Often times a particular choice of sound processing is made, and the resultant parameters from the physical interface are conditioned and mapped to the available sound parameters in an exploratory fashion. The main goal of the work presented here is to demonstrate the importance of the space that lies between physical interface design and the choice of sound manipulation algorithm, and to present a new framework for instrument design that strongly considers this essential part of the design process. In particular, this research takes the viewpoint that instrument designs should be considered in a musical control context, and that both control and sound dynamics must be considered in tandem. In order to achieve this holistic approach, the work presented in this dissertation assumes complementary points of view. Instrument design is first seen as a function of musical context, focusing on electroacoustic music and leading to a view on gesture that relates perceived musical intent to the dynamics of an instrumental system. The important design concept of mapping is then discussed from a theoretical and conceptual point of view, relating perceptual, systems and mathematically-oriented ways of examining the subject. This theoretical framework gives rise to a mapping design space, functional analysis of pertinent existing literature, implementations of mapping tools, instrumental control designs and several perceptual studies that explore the influence of mapping structure. Each of these reflect a high-level approach in which control structures are imposed on top of a high-dimensional space of control and sound synthesis

  16. Radiographic rejection index using statistical process control

    International Nuclear Information System (INIS)

    Savi, M.B.M.B.; Camozzato, T.S.C.; Soares, F.A.P.; Nandi, D.M.

    2015-01-01

    The Repeat Analysis Index (IRR) is one of the items contained in the Quality Control Program dictated by brazilian law of radiological protection and should be performed frequently, at least every six months. In order to extract more and better information of IRR, this study presents the Statistical Quality Control applied to reject rate through Statistical Process Control (Control Chart for Attributes ρ - GC) and the Pareto Chart (GP). Data collection was performed for 9 months and the last four months of collection was given on a daily basis. The Limits of Control (LC) were established and Minitab 16 software used to create the charts. IRR obtained for the period was corresponding to 8.8% ± 2,3% and the generated charts analyzed. Relevant information such as orders for X-ray equipment and processors were crossed to identify the relationship between the points that exceeded the control limits and the state of equipment at the time. The GC demonstrated ability to predict equipment failures, as well as the GP showed clearly what causes are recurrent in IRR. (authors) [pt

  17. Modelling and control of a flotation process

    International Nuclear Information System (INIS)

    Ding, L.; Gustafsson, T.

    1999-01-01

    A general description of a flotation process is given. The dynamic model of a MIMO nonlinear subprocess in flotation, i. e. the pulp levels in five compartments in series is developed and the model is verified with real data from a production plant. In order to reject constant disturbances five extra states are introduced and the model is modified. An exact linearization has been made for the non-linear model and a linear quadratic gaussian controller is proposed based on the linearized model. The simulation result shows an improved performance of the pulp level control when the set points are changed or a disturbance occur. In future the controller will be tested in production. (author)

  18. Methods of control the machining process

    Directory of Open Access Journals (Sweden)

    Yu.V. Petrakov

    2017-12-01

    Full Text Available Presents control methods, differentiated by the time of receipt of information used: a priori, a posteriori and current. When used a priori information to determine the mode of cutting is carried out by simulation the process of cutting allowance, where the shape of the workpiece and the details are presented in the form of wireframes. The office for current information provides for a system of adaptive control and modernization of CNC machine, where in the input of the unit shall be computed by using established optimization software. For the control by a posteriori information of the proposed method of correction of shape-generating trajectory in the second pass measurement surface of the workpiece formed by the first pass. Developed programs that automatically design the adjusted file for machining.

  19. process controller for induction vacuum brazing

    International Nuclear Information System (INIS)

    Aldea, A.

    2016-01-01

    A brazing operation involves joining two parts made of different materials, using a filler material that has a melting temperature lower than the base materials used. The temperature of the process must be carefully controlled, sometimes with an accuracy of about 1°C, because overshooting the prescribed temperature results in detrimental metallurgic phenomena and joints of poor quality. The brazing system is composed of an operating cabinet, a mid-frequency generator, a vacuum chamber with an induction coil inside and the parts that have to be brazed. Until now, to operate this system two operators were required: one to continuously read the temperature with an optical pyrometer and another to manually adjust the current in the induction coil according to his intuition and prediction gained only by experience. The improvement that we made to the system involved creating an automatic temperature control unit, using a PID closed loop controller that reads the temperature of the parts and adjusts automatically the current in the coil. Using the PID controller, the brazing engineer can implement a certain temperature slope for the current brazing process. (authors)

  20. Process management and controlling in diagnostic radiology

    International Nuclear Information System (INIS)

    Gocke, P.; Debatin, J.F.; Duerselen, L.F.J.

    2002-01-01

    Systematic process management and efficient quality control is rapidly gaining importance in our healthcare system. What does this mean for diagnostic radiology departments?To improve efficiency, quality and productivity the workflow within the department of diagnostic and interventional radiology at the University Hospital of Essen were restructured over the last two years. Furthermore, a controlling system was established. One of the pursued aims was to create a quality management system as a basis for the subsequent certification according to the ISO EN 9001:2000 norm.Central to the success of the workflow reorganisation was the training of selected members of the department's staff in process and quality management theory. Thereafter, a dedicated working group was created to prepare the reorganisation and the subsequent ISO certification with the support of a consulting partner. To assure a smooth implementation of the restructured workflow and create acceptance for the required ISO-9001 documentation, the entire staff was familiarized with the basic ideas of process- and quality-management in several training sessions.This manuscript summarizes the basic concepts of process and quality management as they were taught to our staff. A direct relationship towards diagnostic radiology is maintained throughout the text. (orig.) [de

  1. Flotation process control optimisation at Prominent Hill

    International Nuclear Information System (INIS)

    Lombardi, Josephine; Muhamad, Nur; Weidenbach, M.

    2012-01-01

    OZ Minerals' Prominent Hill copper- gold concentrator is located 130 km south east of the town of Coober Pedy in the Gawler Craton of South Australia. The concentrator was built in 2008 and commenced commercial production in early 2009. The Prominent Hill concentrator is comprised of a conventional grinding and flotation processing plant with a 9.6 Mtpa ore throughput capacity. The flotation circuit includes six rougher cells, an IseMill for regrinding the rougher concentrate and a Jameson cell heading up the three stage conventional cell cleaner circuit. In total there are four level controllers in the rougher train and ten level controllers in the cleaning circuit for 18 cells. Generic proportional — integral and derivative (PID) control used on the level controllers alone propagated any disturbances downstream in the circuit that were generated from the grinding circuit, hoppers, between cells and interconnected banks of cells, having a negative impact on plant performance. To better control such disturbances, FloatStar level stabiliser was selected for installation on the flotation circuit to account for the interaction between the cells. Multivariable control was also installed on the five concentrate hoppers to maintain consistent feed to the cells and to the IsaMill. An additional area identified for optimisation in the flotation circuit was the mass pull rate from the rougher cells. FloatStar flow optimiser was selected to be installed subsequent to the FloatStar level stabiliser. This allowed for a unified, consistent and optimal approach to running the rougher circuit. This paper describes the improvement in the stabilisation of the circuit achieved by the FloatStar level stabiliser by using the interaction matrix between cell level controllers and the results and benefits of implementing the FloatStar flow optimiser on the rougher train.

  2. Interference in the processing of adjunct control

    Directory of Open Access Journals (Sweden)

    Dan eParker

    2015-09-01

    Full Text Available Recent research on the memory operations used in language comprehension has revealed a selective profile of interference effects during memory retrieval. Dependencies such as subject-verb agreement show strong facilitatory interference effects from structurally inappropriate but feature-matching distractors, leading to illusions of grammaticality (Dillon, Mishler, Sloggett, & Phillips, 2013; Pearlmutter, Garnsey, & Bock, 1999; Wagers, Lau, & Phillips, 2009. In contrast, dependencies involving reflexive anaphors are generally immune to interference effects (Dillon et al., 2013; Sturt, 2003; Xiang, Dillon, & Phillips, 2009. This contrast has led to the proposal that all anaphors that are subject to structural constraints are immune to facilitatory interference. Here we use an animacy manipulation to examine whether adjunct control dependencies, which involve an interpreted anaphoric relation between a null subject and its licensor, are also immune to facilitatory interference effects. Our results show reliable facilitatory interference in the processing of adjunct control dependencies, which challenges the generalization that anaphoric dependencies as a class are immune to such effects. To account for the contrast between adjunct control and reflexive dependencies, we suggest that variability within anaphora could reflect either an inherent primacy of animacy cues in retrieval processes, or differential degrees of match between potential licensors and the retrieval probe.

  3. Method for enhanced control of welding processes

    Science.gov (United States)

    Sheaffer, Donald A.; Renzi, Ronald F.; Tung, David M.; Schroder, Kevin

    2000-01-01

    Method and system for producing high quality welds in welding processes, in general, and gas tungsten arc (GTA) welding, in particular by controlling weld penetration. Light emitted from a weld pool is collected from the backside of a workpiece by optical means during welding and transmitted to a digital video camera for further processing, after the emitted light is first passed through a short wavelength pass filter to remove infrared radiation. By filtering out the infrared component of the light emitted from the backside weld pool image, the present invention provides for the accurate determination of the weld pool boundary. Data from the digital camera is fed to an imaging board which focuses on a 100.times.100 pixel portion of the image. The board performs a thresholding operation and provides this information to a digital signal processor to compute the backside weld pool dimensions and area. This information is used by a control system, in a dynamic feedback mode, to automatically adjust appropriate parameters of a welding system, such as the welding current, to control weld penetration and thus, create a uniform weld bead and high quality weld.

  4. Monitoring and controlling the biogas process

    Energy Technology Data Exchange (ETDEWEB)

    Ahring, B K; Angelidaki, I [The Technical Univ. of Denmark, Dept. of Environmental Science and Engineering, Lyngby (Denmark)

    1997-08-01

    Many modern large-scale biogas plants have been constructed recently, increasing the demand for proper monitoring and control of these large reactor systems. For monitoring the biogas process, an easy to measure and reliable indicator is required, which reflects the metabolic state and the activity of the bacterial populations in the reactor. In this paper, we discuss existing indicators as well as indicators under development which can potentially be used to monitor the state of the biogas process in a reactor. Furthermore, data are presented from two large scale thermophilic biogas plants, subjected to temperature changes and where the concentration of volatile fatty acids was monitored. The results clearly demonstrated that significant changes in the concentration of the individual VFA occurred although the biogas production was not significantly changed. Especially the concentrations of butyrate, isobutyrate and isovalerate showed significant changes. Future improvements of process control could therefore be based on monitoring of the concentration of specific VFA`s together with information about the bacterial populations in the reactor. The last information could be supplied by the use of modern molecular techniques. (au) 51 refs.

  5. Development of GUS for control applications at the Advanced Photon Source

    International Nuclear Information System (INIS)

    Chung, Y.; Barr, D.; Borland, M.; Kirchman, J.; Decker, G.; Kim, K.

    1994-01-01

    A script-based interpretive shell GUS (General Purpose Data Acquisition for Unix Shell) has been developed for application to the Advanced Photon Source (APS) control. The primary design objective of GUS is to provide a mechanism for efficient data flow among modularized objects called Data Access Modules (DAMs). GUS consists of four major components: user interface, kernel, built-in command module, and DAMS. It also incorporates the Unix shell to make use of the existing utility programs for file manipulation and data analysis. At this time, DAMs have been written for device access through EPICS (Experimental Physics and Industrial Control System), data I/O for SDDS (Self-Describing Data Set) files, matrix manipulation, graphics display, digital signal processing, and beam position feedback system control. The modular and object-oriented construction of GUS will facilitate addition of more DAMs with other functions in the future

  6. A RISC/UNIX workstation second stage trigger

    International Nuclear Information System (INIS)

    Foreman, W.M.; Amann, J.F.; Fu, S.; Kozlowski, T.; Naivar, F.J.; Oothoudt, M.A.; Shelley, F.

    1992-01-01

    Recent advances in Reduced Instruction Set Computer (RISC) workstations have greatly altered the economics of processing power available for experiments. In addition VME interfaces available for many of these workstations make it possible to use them in experiment frontends for filtering and compressing data. Such a second stage trigger has been implemented at LAMPF using a commercially available workstation and VME interface. The implementation is described and measurements of data transfer speeds are presented in this paper

  7. Statistical process control in nursing research.

    Science.gov (United States)

    Polit, Denise F; Chaboyer, Wendy

    2012-02-01

    In intervention studies in which randomization to groups is not possible, researchers typically use quasi-experimental designs. Time series designs are strong quasi-experimental designs but are seldom used, perhaps because of technical and analytic hurdles. Statistical process control (SPC) is an alternative analytic approach to testing hypotheses about intervention effects using data collected over time. SPC, like traditional statistical methods, is a tool for understanding variation and involves the construction of control charts that distinguish between normal, random fluctuations (common cause variation), and statistically significant special cause variation that can result from an innovation. The purpose of this article is to provide an overview of SPC and to illustrate its use in a study of a nursing practice improvement intervention. Copyright © 2011 Wiley Periodicals, Inc.

  8. Multivariate Process Control with Autocorrelated Data

    DEFF Research Database (Denmark)

    Kulahci, Murat

    2011-01-01

    As sensor and computer technology continues to improve, it becomes a normal occurrence that we confront with high dimensional data sets. As in many areas of industrial statistics, this brings forth various challenges in statistical process control and monitoring. This new high dimensional data...... often exhibit not only cross-­‐correlation among the quality characteristics of interest but also serial dependence as a consequence of high sampling frequency and system dynamics. In practice, the most common method of monitoring multivariate data is through what is called the Hotelling’s T2 statistic....... In this paper, we discuss the effect of autocorrelation (when it is ignored) on multivariate control charts based on these methods and provide some practical suggestions and remedies to overcome this problem....

  9. Nanoparticles from a controlled polymerization process

    International Nuclear Information System (INIS)

    Tirumala, V.R.; Caneba, G.T.; Dar, Y.; Wang, H.-H.; Mancini, D.C.

    2003-01-01

    Free-radical retrograde precipitation polymerization process in the past has shown excellent control characteristics over reaction rate, molecular weight, and in the entrapment of live radicals for the generation of block copolymers. The same principle has now been extended to study the reaction confinement to a nanoscale region. Nanosized polymer particles have been reported to form from block copolymers, conventional precipitation polymerization methods, or through emulsion polymerization approaches. In this work, we present a new method of generating nanosized polymer particles by polymerizing the monomer in an environment that precipitates the polymer above the lower critical solution temperature. The nanoparticles have been characterized by both tapping-mode atomic force microscopy observations and in situ synchrotron time-resolved small-angle X-ray scattering analysis. The results from both the techniques showed the formation of nanoparticles in the size range of 15-30 nm, directly from the polymerization process.

  10. Thermodynamics, kinetics and process control of nitriding

    DEFF Research Database (Denmark)

    Mittemeijer, Eric J.; Somers, Marcel A. J.

    1997-01-01

    As a prerequisite for the predictability of properties obtained by a nitriding treatment of iron based workpieces, the relation between the process parameters and the composition and structure of the surface layer produced must be known. At present, even the description of thermodynamic equilibrium...... of pure Fe-N phases has not been fully achieved. It is shown that taking into account the ordering of nitrogen in the epsilon and gamma' iron nitride phases leads to an improved understanding of the Fe-N phase diagram. Although consideration of thermodynamics indicates the state the system strives for...... for process control of gaseous nitriding by monitoring the partial pressure of oxygen in the furnace using a solid state electrolyte is provided. At the time the work was carried out the authors were in the Laboratory of Materials Science, Delft University of Technology, Rotterdamseweg 137, 2628 AL Delft...

  11. High risk process control system assessment methodology

    Energy Technology Data Exchange (ETDEWEB)

    Santos, Venetia [Pontificia Universidade Catolica do Rio de Janeiro (PUC-Rio), RJ (Brazil); Zamberlan, Maria Cristina [National Institute of Tehnology (INT), Rio de Janeiro, RJ (Brazil). Human Reliability and Ergonomics Research Group for the Oil, Gas and Energy Sector

    2009-07-01

    The evolution of ergonomics methodology has become necessary due to the dynamics imposed by the work environment, by the increase of the need of human cooperation and by the high interaction between various sections within a company. In the last 25 years, as of studies made in the high risk process control, we have developed a methodology to evaluate these situations that focus on the assessment of activities and human cooperation, the assessment of context, the assessment of the impact of work of other sectors in the final activity of the operator, as well as the modeling of existing risks. (author)

  12. Transforming the Duke Power work control process

    International Nuclear Information System (INIS)

    Dulin, R.

    1996-01-01

    Faced with rising operating and maintenance costs, Duke Power initiated a Work Control Project to provide fundamental rethinking, dramatic quality improvements, and a dramatic reduction in inefficiencies. Other aims were: to do more better with less, to improve coordination between work groups, reduce paperwork, increase effectiveness and utilization of station personnel, and achieve consistent implementation between sites. The existing electronic work management scheme needed some modifications to its software, especially, the programming of a screen to allow simple entry of corrective problems, and the implementation of the new scheduling process. The project has been successful in speeding up the resolution of problems, and in reducing backlogs of maintenance work

  13. Method of controlling radioactive waste processing systems

    International Nuclear Information System (INIS)

    Mikawa, Hiroji; Sato, Takao.

    1981-01-01

    Purpose: To minimize the pellet production amount, maximize the working life of a solidifying device and maintaining the mechanical strength of pellets to a predetermined value irrespective of the type and the cycle of occurrence of the secondary waste in the secondary waste solidifying device for radioactive waste processing systems in nuclear power plants. Method: Forecasting periods for the type, production amount and radioactivity level of the secondary wastes are determined in input/output devices connected to a control system and resulted signals are sent to computing elements. The computing elements forecast the production amount of regenerated liquid wastes after predetermined days based on the running conditions of a condensate desalter and the production amounts of filter sludges and liquid resin wastes after predetermined days based on the liquid waste processing amount or the like in a processing device respectively. Then, the mass balance between the type and the amount of the secondary wastes presently stored in a tank are calculated and the composition and concentration for the processing liquid are set so as to obtain predetermined values for the strength of pellets that can be dried to solidify, the working life of the solidifying device itself and the radioactivity level of the pellets. Thereafter, the running conditions for the solidifying device are determined so as to maximize the working life of the solidifying device. (Horiuchi, T.)

  14. Data processing software for purex plant process control laboratory

    International Nuclear Information System (INIS)

    Kansara, V.P.; Achuthan, P.V.; Sridhar, S.; Ramanujam, A.; Dhumwad, R.K.

    1990-01-01

    A software has been developed at the Fuel Reprocessing Division, Trombay to meet the data processing needs of the Control Laboratory of a reprocessing plant. During the normal plant operations contents of over one hundred process tanks have to be sampled and analysed for regular monitoring. In order to speed up the computation and the reporting of results as well as to obtain the process performance data over a period of time a software has been developed. The package has been sucessfully demonstrated and implemented at the Plutonium Plant, Trombay. This has been in continuous use since May 1987 with highly satisfactory performance. The software is a totally menu-driven package which can be used by the laboratory analysts with a few hours of training. The features include data validation involving source tank identification, the nature of the sample, the range of expected results, any duplication in sample numbering etc. Audio indication of deviations from the expected input or output values are given with an option to override in case of abnormal samples. The progress of analysis can be obtained for a given sample at any given time. Incorporated in the software is the help menu for quick reference of analytical protocol to be followed for a given tank/method. The computations for the determinations are carried out after obtaining input values on a screen-form. Th e results can be displayed on the monitor or obtained in the form of a hard copy i n any desired format. (author). 17 figs., 2 refs

  15. Control system for technological processes in tritium processing plants with process analysis

    International Nuclear Information System (INIS)

    Retevoi, Carmen Maria; Stefan, Iuliana; Balteanu, Ovidiu; Stefan, Liviu; Bucur, Ciprian

    2005-01-01

    Integration of a large variety of installations and equipment into a unitary system for controlling the technological process in tritium processing nuclear facilities appears to be a rather complex approach particularly when experimental or new technologies are developed. Ensuring a high degree of versatility allowing easy modifications in configurations and process parameters is a major requirement imposed on experimental installations. The large amount of data which must be processed, stored and easily accessed for subsequent analyses imposes development of a large information network based on a highly integrated system containing the acquisition, control and technological process analysis data as well as data base system. On such a basis integrated systems of computation and control able to conduct the technological process could be developed as well protection systems for cases of failures or break down. The integrated system responds to the control and security requirements in case of emergency and of the technological processes specific to the industry that processes radioactive or toxic substances with severe consequences in case of technological failure as in the case of tritium processing nuclear plant. In order to lower the risk technological failure of these processes an integrated software, data base and process analysis system are developed, which, based on identification algorithm of the important parameters for protection and security systems, will display the process evolution trend. The system was checked on a existing plant that includes a removal tritium unit, finally used in a nuclear power plant, by simulating the failure events as well as the process. The system will also include a complete data base monitoring all the parameters and a process analysis software for the main modules of the tritium processing plant, namely, isotope separation, catalytic purification and cryogenic distillation

  16. Hybrid scatterometry measurement for BEOL process control

    Science.gov (United States)

    Timoney, Padraig; Vaid, Alok; Kang, Byeong Cheol; Liu, Haibo; Isbester, Paul; Cheng, Marjorie; Ng-Emans, Susan; Yellai, Naren; Sendelbach, Matt; Koret, Roy; Gedalia, Oram

    2017-03-01

    Scaling of interconnect design rules in advanced nodes has been accompanied by a reducing metrology budget for BEOL process control. Traditional inline optical metrology measurements of BEOL processes rely on 1-dimensional (1D) film pads to characterize film thickness. Such pads are designed on the assumption that solid copper blocks from previous metallization layers prevent any light from penetrating through the copper, thus simplifying the effective film stack for the 1D optical model. However, the reduction of the copper thickness in each metallization layer and CMP dishing effects within the pad, have introduced undesired noise in the measurement. To resolve this challenge and to measure structures that are more representative of product, scatterometry has been proposed as an alternative measurement. Scatterometry is a diffraction based optical measurement technique using Rigorous Coupled Wave Analysis (RCWA), where light diffracted from a periodic structure is used to characterize the profile. Scatterometry measurements on 3D structures have been shown to demonstrate strong correlation to electrical resistance parameters for BEOL Etch and CMP processes. However, there is significant modeling complexity in such 3D scatterometry models, in particlar due to complexity of front-end-of-line (FEOL) and middle-of-line (MOL) structures. The accompanying measurement noise associated with such structures can contribute significant measurement error. To address the measurement noise of the 3D structures and the impact of incoming process variation, a hybrid scatterometry technique is proposed that utilizes key information from the structure to significantly reduce the measurement uncertainty of the scatterometry measurement. Hybrid metrology combines measurements from two or more metrology techniques to enable or improve the measurement of a critical parameter. In this work, the hybrid scatterometry technique is evaluated for 7nm and 14nm node BEOL measurements of

  17. Traffic analysis and control using image processing

    Science.gov (United States)

    Senthilkumar, K.; Ellappan, Vijayan; Arun, A. R.

    2017-11-01

    This paper shows the work on traffic analysis and control till date. It shows an approach to regulate traffic the use of image processing and MATLAB systems. This concept uses computational images that are to be compared with original images of the street taken in order to determine the traffic level percentage and set the timing for the traffic signal accordingly which are used to reduce the traffic stoppage on traffic lights. They concept proposes to solve real life scenarios in the streets, thus enriching the traffic lights by adding image receivers like HD cameras and image processors. The input is then imported into MATLAB to be used. as a method for calculating the traffic on roads. Their results would be computed in order to adjust the traffic light timings on a particular street, and also with respect to other similar proposals but with the added value of solving a real, big instance.

  18. AN OVERVIEW OF PHARMACEUTICAL PROCESS VALIDATION AND PROCESS CONTROL VARIABLES OF TABLETS MANUFACTURING PROCESSES IN INDUSTRY

    OpenAIRE

    Mahesh B. Wazade*, Sheelpriya R. Walde and Abhay M. Ittadwar

    2012-01-01

    Validation is an integral part of quality assurance; the product quality is derived from careful attention to a number of factors including selection of quality parts and materials, adequate product and manufacturing process design, control of the process variables, in-process and end-product testing. Recently validation has become one of the pharmaceutical industry’s most recognized and discussed subjects. It is a critical success factor in product approval and ongoing commercialization, fac...

  19. Bayesian nonparametric adaptive control using Gaussian processes.

    Science.gov (United States)

    Chowdhary, Girish; Kingravi, Hassan A; How, Jonathan P; Vela, Patricio A

    2015-03-01

    Most current model reference adaptive control (MRAC) methods rely on parametric adaptive elements, in which the number of parameters of the adaptive element are fixed a priori, often through expert judgment. An example of such an adaptive element is radial basis function networks (RBFNs), with RBF centers preallocated based on the expected operating domain. If the system operates outside of the expected operating domain, this adaptive element can become noneffective in capturing and canceling the uncertainty, thus rendering the adaptive controller only semiglobal in nature. This paper investigates a Gaussian process-based Bayesian MRAC architecture (GP-MRAC), which leverages the power and flexibility of GP Bayesian nonparametric models of uncertainty. The GP-MRAC does not require the centers to be preallocated, can inherently handle measurement noise, and enables MRAC to handle a broader set of uncertainties, including those that are defined as distributions over functions. We use stochastic stability arguments to show that GP-MRAC guarantees good closed-loop performance with no prior domain knowledge of the uncertainty. Online implementable GP inference methods are compared in numerical simulations against RBFN-MRAC with preallocated centers and are shown to provide better tracking and improved long-term learning.

  20. DB2 9 for Linux, Unix, and Windows database administration certification study guide

    CERN Document Server

    Sanders, Roger E

    2007-01-01

    In DB2 9 for Linux, UNIX, and Windows Database Administration Certification Study Guide, Roger E. Sanders-one of the world's leading DB2 authors and an active participant in the development of IBM's DB2 certification exams-covers everything a reader needs to know to pass the DB2 9 UDB DBA Certification Test (731).This comprehensive study guide steps you through all of the topics that are covered on the test, including server management, data placement, database access, analyzing DB2 activity, DB2 utilities, high availability, security, and much more. Each chapter contains an extensive set of p

  1. FORTRAN data files transference from VAX/VMS to ALPHA/UNIX

    International Nuclear Information System (INIS)

    Sanchez, E.; Milligen, B.Ph. van

    1997-01-01

    Several tools have been developed to access the TJ-I and TJ-IU databases, which currently reside in VAX/VMS servers, from the TJ-II Data Acquisition System DEC ALPHA 8400 server. The TJ-I/TJ-IU databases are not homogeneous and contain several types of data files, namely, SADE. CAMAC and FORTRAN un formatted files. The tools presented in this report allow one to transfer CAMAC and those FORTRAN un formatted files defined herein. from a VAX/VMS server, for data manipulation on the ALPHA/Digital UNIX server. (Author) 5 refs

  2. Security of legacy process control systems : Moving towards secure process control systems

    NARCIS (Netherlands)

    Oosterink, M.

    2012-01-01

    This white paper describes solutions which organisations may use to improve the security of their legacy process control systems. When we refer to a legacy system, we generally refer to old methodologies, technologies, computer systems or applications which are still in use, despite the fact that

  3. Practical Implementations of Advanced Process Control for Linear Systems

    DEFF Research Database (Denmark)

    Knudsen, Jørgen K . H.; Huusom, Jakob Kjøbsted; Jørgensen, John Bagterp

    2013-01-01

    This paper describes some practical problems encountered, when implementing Advanced Process Control, APC, schemes on linear processes. The implemented APC controllers discussed will be LQR, Riccati MPC and Condensed MPC controllers illustrated by simulation of the Four Tank Process and a lineari......This paper describes some practical problems encountered, when implementing Advanced Process Control, APC, schemes on linear processes. The implemented APC controllers discussed will be LQR, Riccati MPC and Condensed MPC controllers illustrated by simulation of the Four Tank Process...... on pilot plant equipment on the department of Chemical Engineering DTU Lyngby....

  4. Neuro-fuzzy Control of Integrating Processes

    Directory of Open Access Journals (Sweden)

    Anna Vasičkaninová

    2011-11-01

    Full Text Available Fuzzy technology is adaptive and easily applicable in different areas.Fuzzy logic provides powerful tools to capture the perceptionof natural phenomena. The paper deals with tuning of neuro-fuzzy controllers for integrating plant and for integrating plantswith time delay. The designed approach is verified on three examples by simulations and compared plants with classical PID control.Designed fuzzy controllers lead to better closed-loop control responses then classical PID controllers.

  5. On-line process control monitoring system

    International Nuclear Information System (INIS)

    O'Rourke, P.E.; Van Hare, D.R.; Prather, W.S.

    1992-01-01

    This patent describes apparatus for monitoring at a plurality of locations within a system the concentration of at least one chemical substance involved in a chemical process. It comprises plurality of process cells; first means for carrying the light; second means for carrying the light; means for producing a spectrum from the light received by the second carrying means; multiplexing means for selecting one process cell of the plurality of process cells at a time so that the producing means can produce a process spectrum from the one cell of the process cells; a reference cell for producing a reference spectrum for comparison to the process spectrum; a standard cell for producing a standard spectrum for comparison to the process spectrum; and means for comparing the reference spectrum, the standard spectrum and the process spectrum and determining the concentration of the chemical substance in the process cell

  6. Optical metrology for advanced process control: full module metrology solutions

    Science.gov (United States)

    Bozdog, Cornel; Turovets, Igor

    2016-03-01

    Optical metrology is the workhorse metrology in manufacturing and key enabler to patterning process control. Recent advances in device architecture are gradually shifting the need for process control from the lithography module to other patterning processes (etch, trim, clean, LER/LWR treatments, etc..). Complex multi-patterning integration solutions, where the final pattern is the result of multiple process steps require a step-by-step holistic process control and a uniformly accurate holistic metrology solution for pattern transfer for the entire module. For effective process control, more process "knobs" are needed, and a tighter integration of metrology with process architecture.

  7. Comparison Analysis of Model Predictive Controller with Classical PID Controller For pH Control Process

    Directory of Open Access Journals (Sweden)

    V. Balaji

    2016-12-01

    Full Text Available pH control plays a important role in any chemical plant and process industries. For the past four decades the classical PID controller has been occupied by the industries. Due to the faster computing   technology in the industry demands a tighter advanced control strategy. To fulfill the needs and requirements Model Predictive Control (MPC is the best among all the advanced control algorithms available in the present scenario. The study and analysis has been done for First Order plus Delay Time (FOPDT model controlled by Proportional Integral Derivative (PID and MPC using the Matlab software. This paper explores the capability of the MPC strategy, analyze and compare the control effects with conventional control strategy in pH control. A comparison results between the PID and MPC is plotted using the software. The results clearly show that MPC provide better performance than the classical controller.

  8. Anticipated Activities in Maritime Work, Process Control, and Business Processes

    DEFF Research Database (Denmark)

    Andersen, Peter Bøgh

    2004-01-01

    Most activities are anticipated before they are executed. The paper presents methods for describing this anticipated state and the processes that may lead to a new state where the activities are executed. The method builds on linguistic case-theory.......Most activities are anticipated before they are executed. The paper presents methods for describing this anticipated state and the processes that may lead to a new state where the activities are executed. The method builds on linguistic case-theory....

  9. Using the Shewhart control charts by process control

    Directory of Open Access Journals (Sweden)

    Katarína Lestyánszka Škůrková

    2013-04-01

    Full Text Available This article deals with the statistics pursing the process capability of turning in screws production in RIBE Slovakia, k.s. in technical practice, an important group of statistic methods is formed by analyzing qualification of measures, production equipment and qualification of process. By the term “process qualification“ we mean the ability of the process to observe required technical parameters by required value and tolerance limits. Findings of the process capability can be isolated in the estimate process capability (before starting the production and permanent process capability. Also, we have finished the quest for the process capability where the indexes Cp and Cpk are bigger than the determined value of 1.33 points.

  10. Modeling, simulation and control of fermentation processes

    OpenAIRE

    Aros, Nelson; Cifuentes, Marcelo; Mardones, Javier

    2011-01-01

    Se presenta un simulador en ambiente Matlab/Simulink para controlar el proceso de fermentación, diseñando una ley de control invariante, luego un control estabilizante para mantener la planta dentro del sistema de referencia y finalmente un control estabilizante mejorado con la implementación de un controlador difuso. El modelo Monod es utilizado y vía simulación se prueban las estrategias aludidas para controlar la tasa de crecimiento específica de la biomasa. El control estabilizante aplica...

  11. An Overview of Pharmaceutical Validation and Process Controls in ...

    African Journals Online (AJOL)

    It has always been known that the processes involved in pharmaceutical production impact significantly on the quality of the products The processes include raw material and equipment inspections as well as in-process controls. Process controls are mandatory in good manufacturing practice (GMP). The purpose is to ...

  12. Verification of product quality from process control

    International Nuclear Information System (INIS)

    Drobot, A.; Bunnell, L.R.; Freeborn, W.P.; Macedo, P.B.; Mellinger, G.B.; Pegg, I.L.; Piepel, G.F.; Reimus, M.A.H.; Routt, K.R.; Saad, E.

    1989-01-01

    Process models were developed to characterize the waste vitrification at West Valley, in terms of process operating constraints and glass compositions achievable. The need for verification of compliance with the proposed Waste Acceptance Preliminary Specification criteria led to development of product models, the most critical one being a glass durability model. Both process and product models were used in developing a target composition for the waste glass. This target composition designed to ensure that glasses made to this target will be of acceptable durability after all process variations have been accounted for. 4 refs., 11 figs., 5 tabs

  13. Fluidization control in the wurster coating process

    Directory of Open Access Journals (Sweden)

    el Mafadi Samira

    2003-01-01

    Full Text Available Paniculate coating process in a fluidized bed involves different sub processes including particle wetting, spreading and also consolidation or drying of the coating applied. These sub processes are done simultaneously to particle fluidization and motion. All the parameters of fluidization are known to affect the coating quality. That is why the motion of particles in the Wurster coating process has been observed and described step by step. These observations have achieved a general understanding of phenomena which take place inside the bed during fluidization and have allowed the development of an easy method for optimizing all the parameters affecting this operation.

  14. Association between product quality control and process quality control of bulk milk

    NARCIS (Netherlands)

    Velthuis, A.; Asseldonk, van M.A.P.M.

    2010-01-01

    Assessment of dairy-milk quality is based on product quality control (testing bulk-milk samples) and process quality control (auditing dairy farms). It is unknown whether process control improves product quality. To quantify possible association between product control and process control a

  15. Plug-and-play process control

    DEFF Research Database (Denmark)

    Bendtsen, Jan Dimon; Trangbæk, K; Stoustrup, Jakob

    2008-01-01

    An important issue in the area of reconfigurable systems is how to respond correctly if new components are added. We consider the problem of improving control performance for a system where a new set of sensors becomes available. It is assumed that a complete re-design of the control system is un...

  16. Overview of advanced process control in welding within ERDA

    International Nuclear Information System (INIS)

    Armstrong, R.E.

    1977-01-01

    The special kinds of demands placed on ERDA weapons and reactors require them to have very reliable welds. Process control is critical in achieving this reliability. ERDA has a number of advanced process control projects underway with much of the emphasis being on electron beam welding. These include projects on voltage measurement, beam-current control, beam focusing, beam spot tracking, spike suppression, and computer control. A general discussion of process control in welding is followed by specific examples of some of the advanced joining process control projects in ERDA

  17. Intelligent control for scalable video processing

    NARCIS (Netherlands)

    Wüst, C.C.

    2006-01-01

    In this thesis we study a problem related to cost-effective video processing in software by consumer electronics devices, such as digital TVs. Video processing is the task of transforming an input video signal into an output video signal, for example to improve the quality of the signal. This

  18. Viscous Control of the Foam Glass Process

    DEFF Research Database (Denmark)

    Petersen, Rasmus Rosenlund; König, Jakob; Smedskjær, Morten Mattrup

    The production of foam glass as heat insulating material is an important industrial process because it enables low-cost recycling of glass waste from a variety of chemical compositions. Optimization of the foaming process of new glass waste compositions is time consuming, since many factors affect...... the foaming process such as temperature, particle size, type and concentration of foaming agent. The foaming temperature is one of the key factors, because even small temperature changes can affect the melt viscosity by several orders of magnitude. Therefore, it is important to establish the viscosity range...... in which the foaming process should take place, particularly when the type of recycled cullet is changed or several types of cullet are mixed in one batch. According to recent glass literature, the foaming process should occur at viscosity 103 to 105 Pa s. However, no systematic studies have hitherto been...

  19. The COSY control system, a distributed realtime operating system: First practical experience at the COSY-injector

    International Nuclear Information System (INIS)

    Stephan, M.; Hacker, U.; Henn, K.; Richert, A.; Sobotta, K.; Weinert, A.

    1991-01-01

    The COSY control system is hierarchically organized with distributed intelligence and autonomous processing units for dedicated components. Data communication is performed via LAN and over a fieldbus. The hostsystems are UNIX-based, whereas the field-controllers are running a modular realtime operating-system RT/OS which has been developed at KFA. The computer-hardware consists of RISC mini computers, VME-computers in the field and G64 equipment-control-module in geographical expansion of the controller by a fieldbus based on the PDV-standard. The man-machine interface consists of X-window based work stations. On top of X-window a graphical user interface based on object oriented methods is used. A distributed realtime data base allows access to the accelerator state from every workstation. A special highlevel language debugger hosted on the UNIX based workstation and connected over LAN to the VME targets will be used. Together with the software development system for UNIX applications an uniform view of the system appears to the programmer. First practical experience at the COSY injector is presented

  20. Functional Dual Adaptive Control with Recursive Gaussian Process Model

    International Nuclear Information System (INIS)

    Prüher, Jakub; Král, Ladislav

    2015-01-01

    The paper deals with dual adaptive control problem, where the functional uncertainties in the system description are modelled by a non-parametric Gaussian process regression model. Current approaches to adaptive control based on Gaussian process models are severely limited in their practical applicability, because the model is re-adjusted using all the currently available data, which keeps growing with every time step. We propose the use of recursive Gaussian process regression algorithm for significant reduction in computational requirements, thus bringing the Gaussian process-based adaptive controllers closer to their practical applicability. In this work, we design a bi-criterial dual controller based on recursive Gaussian process model for discrete-time stochastic dynamic systems given in an affine-in-control form. Using Monte Carlo simulations, we show that the proposed controller achieves comparable performance with the full Gaussian process-based controller in terms of control quality while keeping the computational demands bounded. (paper)

  1. Deposit control in process cooling water systems

    International Nuclear Information System (INIS)

    Venkataramani, B.

    1981-01-01

    In order to achieve efficient heat transfer in cooling water systems, it is essential to control the fouling of heat exchanger surfaces. Solubilities of scale forming salts, their growth into crystals, and the nature of the surfaces play important roles in the deposition phenomenon. Condensed phosphates, organic polymers and compounds like phosphates are effective in controlling deposition of scale forming salts. The surface active agents inhibit crystal growth and modify the crystals of the scale forming salts, and thus prevent deposition of dense, uniformly structured crystalline mass on the heat transfer surface. Understanding the mechanism of biofouling is essential to control it by surface active agents. Certain measures taken in the plant, such as back flushing, to control scaling, sometimes may not be effective and can be detrimental to the system itself. (author)

  2. UNIX veida mikrokodola operētājsistēma FPGA procesoram

    OpenAIRE

    Liepkalns, Ansis

    2012-01-01

    Risinājumos, kuros izmanto programmējamo loģisko mezglu masīvu (FPGA) procesorus, programmatūras pieejamība ir svarīga, lai samazinātu galaprodukta iegūšanai nepieciešamo laiku. Plašu programmatūras atbalstu ir ieguvušas UNIX veida operētājsistēmas. To kombinācija ar FPGA procesoriem spēj nodrošināt vēlamo izstrādes ātrumu. Lai apmierinātu kvalitātes prasības, tiek piedāvāts izmantot mikrokodola operētājsistēmu. Darbā tiek apskatīta sistēmas mikroshēmā izveide darbībai ar „Minix 3“ mikrokodol...

  3. DB2 9 for Linux, Unix, and Windows database administration upgrade certification study guide

    CERN Document Server

    Sanders, Roger E

    2007-01-01

    Written by one of the world's leading DB2 authors who is an active participant in the development of the DB2 certification exams, this resource covers everything a database adminstrator needs to know to pass the DB2 9 for Linux, UNIX, and Windows Database Administration Certification Upgrade exam (Exam 736). This comprehensive study guide discusses all exam topics: server management, data placement, XML concepts, analyzing activity, high availability, database security, and much more. Each chapter contains an extensive set of practice questions along with carefully explained answers. Both information-technology professionals who have experience as database administrators and have a current DBA certification on version 8 of DB2 and individuals who would like to learn the new features of DB2 9 will benefit from the information in this reference guide.

  4. Screen-based process control in nuclear plants

    International Nuclear Information System (INIS)

    Hinz, W.; Arnoldt, C.; Hessler, C.

    1993-01-01

    Requirements, development and conceptual design of a screen-based control room for nuclear power plants are outlined. The control room consists of three or four equally equipped operator workstations comprising screens for process information and manual process control. A plant overview will assist the coordination among the operators. A safety classified backup system (safety control area) is provided to cover postulated failures of the control means. Some aspects of ergonomical validation and of future development trends are discussed. (orig.) [de

  5. Process control for the ISABELLE system

    International Nuclear Information System (INIS)

    Frankel, R.; Herman, A.; Stattel, P.; Warkentien, R.

    1981-01-01

    A low cost, highly reliable control system has been developed for use in the radiation environment of the ISABELLE ring. The Ultra-High Vacuum installation will consist of approximately 1500 stations - each having 8 readbacks, 4 on-off operations and 4 special functions - distributed over 2.5 mills. The paper will describe a multidrop party line system in which a μp based controller located in a protected environment communicates with up to 25 vacuum stations. All the μp controllers will be linked via the ISABELLE data highway. Mechanical, electronic and software aspects of the system are discussed and performance data from our First Cell model are presented

  6. NONLINEAR MODEL PREDICTIVE CONTROL OF CHEMICAL PROCESSES

    Directory of Open Access Journals (Sweden)

    SILVA R. G.

    1999-01-01

    Full Text Available A new algorithm for model predictive control is presented. The algorithm utilizes a simultaneous solution and optimization strategy to solve the model's differential equations. The equations are discretized by equidistant collocation, and along with the algebraic model equations are included as constraints in a nonlinear programming (NLP problem. This algorithm is compared with the algorithm that uses orthogonal collocation on finite elements. The equidistant collocation algorithm results in simpler equations, providing a decrease in computation time for the control moves. Simulation results are presented and show a satisfactory performance of this algorithm.

  7. Electrolyte chemistry control in electrodialysis processing

    Science.gov (United States)

    Hayes, Thomas D.; Severin, Blaine F.

    2017-12-26

    Methods for controlling electrolyte chemistry in electrodialysis units having an anode and a cathode each in an electrolyte of a selected concentration and a membrane stack disposed therebetween. The membrane stack includes pairs of cationic selective and anionic membranes to segregate increasingly dilute salts streams from concentrated salts stream. Electrolyte chemistry control is via use of at least one of following techniques: a single calcium exclusionary cationic selective membrane at a cathode cell boundary, an exclusionary membrane configured as a hydraulically isolated scavenger cell, a multivalent scavenger co-electrolyte and combinations thereof.

  8. The application of mean control chart in managing industrial processes

    Directory of Open Access Journals (Sweden)

    Papić-Blagojević Nataša

    2013-01-01

    Full Text Available Along with the advent of mass production comes the problem of monitoring and maintaining the quality of the product, which stressed the need for the application of selected statistical and mathematical methods in the control process. The main objective of applying the methods of statistical control is continuous quality improvement through permanent monitoring of the process in order to discover the causes of errors. Shewart charts are the most popular method of statistical process control, which performs separation of controlled and uncontrolled variations along with detection of increased variations. This paper presents the example of Shewart mean control chart with application in managing industrial process.

  9. Automatic process control in anaerobic digestion technology: A critical review.

    Science.gov (United States)

    Nguyen, Duc; Gadhamshetty, Venkataramana; Nitayavardhana, Saoharit; Khanal, Samir Kumar

    2015-10-01

    Anaerobic digestion (AD) is a mature technology that relies upon a synergistic effort of a diverse group of microbial communities for metabolizing diverse organic substrates. However, AD is highly sensitive to process disturbances, and thus it is advantageous to use online monitoring and process control techniques to efficiently operate AD process. A range of electrochemical, chromatographic and spectroscopic devices can be deployed for on-line monitoring and control of the AD process. While complexity of the control strategy ranges from a feedback control to advanced control systems, there are some debates on implementation of advanced instrumentations or advanced control strategies. Centralized AD plants could be the answer for the applications of progressive automatic control field. This article provides a critical overview of the available automatic control technologies that can be implemented in AD processes at different scales. Copyright © 2015 Elsevier Ltd. All rights reserved.

  10. Process plant equipment operation, control, and reliability

    CERN Document Server

    Holloway, Michael D; Onyewuenyi, Oliver A

    2012-01-01

    "Process Plant Equipment Book is another great publication from Wiley as a reference book for final year students as well as those who will work or are working in chemical production plants and refinery…" -Associate Prof. Dr. Ramli Mat, Deputy Dean (Academic), Faculty of Chemical Engineering, Universiti Teknologi Malaysia "…give[s] readers access to both fundamental information on process plant equipment and to practical ideas, best practices and experiences of highly successful engineers from around the world… The book is illustrated throughout with numerous black & white p

  11. Science-based information processing in the process control of power stations

    International Nuclear Information System (INIS)

    Weisang, C.

    1992-01-01

    Through the application of specialized systems, future-orientated information processing integrates the sciences of processes, control systems, process control strategies, user behaviour and ergonomics. Improvements in process control can be attained, inter alia, by the preparation of the information contained (e.g. by suppressing the flow of signals and replacing it with signals which are found on substance) and also by an ergonomic representation of the study of the process. (orig.) [de

  12. Process control system for fresh concrete preparation

    International Nuclear Information System (INIS)

    Bachvarov, N.; Pavlov, P.; Shukov, H.

    1983-01-01

    The paper discusses the principles, structure and organization of a modular microprocessor based control system, designed to be used in fresh concrete fabrication plants. The system is based on the measurement of the aggregate moisture by means of a neutron moisture meter. (author)

  13. Optimal control of raw timber production processes

    Science.gov (United States)

    Ivan Kolenka

    1978-01-01

    This paper demonstrates the possibility of optimal planning and control of timber harvesting activ-ities with mathematical optimization models. The separate phases of timber harvesting are represented by coordinated models which can be used to select the optimal decision for the execution of any given phase. The models form a system whose components are connected and...

  14. Robot welding process control development task

    Science.gov (United States)

    Romine, Peter L.

    1992-01-01

    The completion of, and improvements made to, the software developed during 1990 for program maintenance on the PC and HEURIKON and transfer to the CYRO, and integration of the Rocketdyne vision software with the CYRO is documented. The new programs were used successfully by NASA, Rocketdyne, and UAH technicians and engineers to create, modify, upload, download, and control CYRO NC programs.

  15. Biomolecular Modeling in a Process Dynamics and Control Course

    Science.gov (United States)

    Gray, Jeffrey J.

    2006-01-01

    I present modifications to the traditional course entitled, "Process dynamics and control," which I renamed "Modeling, dynamics, and control of chemical and biological processes." Additions include the central dogma of biology, pharmacokinetic systems, population balances, control of gene transcription, and large­-scale…

  16. Applying Statistical Process Control to Clinical Data: An Illustration.

    Science.gov (United States)

    Pfadt, Al; And Others

    1992-01-01

    Principles of statistical process control are applied to a clinical setting through the use of control charts to detect changes, as part of treatment planning and clinical decision-making processes. The logic of control chart analysis is derived from principles of statistical inference. Sample charts offer examples of evaluating baselines and…

  17. Multi-Model Adaptive Fuzzy Controller for a CSTR Process

    Directory of Open Access Journals (Sweden)

    Shubham Gogoria

    2015-09-01

    Full Text Available Continuous Stirred Tank Reactors are intensively used to control exothermic reactions in chemical industries. It is a very complex multi-variable system with non-linear characteristics. This paper deals with linearization of the mathematical model of a CSTR Process. Multi model adaptive fuzzy controller has been designed to control the reactor concentration and temperature of CSTR process. This method combines the output of multiple Fuzzy controllers, which are operated at various operating points. The proposed solution is a straightforward implementation of Fuzzy controller with gain scheduler to control the linearly inseparable parameters of a highly non-linear process.

  18. 21 CFR 820.70 - Production and process controls.

    Science.gov (United States)

    2010-04-01

    ...) MEDICAL DEVICES QUALITY SYSTEM REGULATION Production and Process Controls § 820.70 Production and process... used as part of production or the quality system, the manufacturer shall validate computer software for... 21 Food and Drugs 8 2010-04-01 2010-04-01 false Production and process controls. 820.70 Section...

  19. New Principles of Process Control in Geotechnics by Acoustic Methods

    OpenAIRE

    Leššo, I.; Flegner, P.; Pandula, B.; Horovčák, P.

    2007-01-01

    The contribution describes the new solution of the control of rotary drilling process as some elementary process in geotechnics. The article presents the first results of research on the utilization of acoustic methods in identification process by optimal control of rotary drilling.

  20. New Principles of Process Control in Geotechnics by Acoustic Methods

    Directory of Open Access Journals (Sweden)

    Leššo, I.

    2007-01-01

    Full Text Available The contribution describes the new solution of the control of rotary drilling process as some elementary process in geotechnics. The article presents the first results of research on the utilization of acoustic methods in identification process by optimal control of rotary drilling.

  1. Big Data Analytics for Industrial Process Control

    DEFF Research Database (Denmark)

    Khan, Abdul Rauf; Schioler, Henrik; Kulahci, Murat

    2017-01-01

    Today, in modern factories, each step in manufacturing produces a bulk of valuable as well as highly precise information. This provides a great opportunity for understanding the hidden statistical dependencies in the process. Systematic analysis and utilization of advanced analytical methods can...

  2. Modeling of Reaction Processes Controlled by Diffusion

    International Nuclear Information System (INIS)

    Revelli, Jorge

    2003-01-01

    Stochastic modeling is quite powerful in science and technology.The technics derived from this process have been used with great success in laser theory, biological systems and chemical reactions.Besides, they provide a theoretical framework for the analysis of experimental results on the field of particle's diffusion in ordered and disordered materials.In this work we analyze transport processes in one-dimensional fluctuating media, which are media that change their state in time.This fact induces changes in the movements of the particles giving rise to different phenomena and dynamics that will be described and analyzed in this work.We present some random walk models to describe these fluctuating media.These models include state transitions governed by different dynamical processes.We also analyze the trapping problem in a lattice by means of a simple model which predicts a resonance-like phenomenon.Also we study effective diffusion processes over surfaces due to random walks in the bulk.We consider different boundary conditions and transitions movements.We derive expressions that describe diffusion behaviors constrained to bulk restrictions and the dynamic of the particles.Finally it is important to mention that the theoretical results obtained from the models proposed in this work are compared with Monte Carlo simulations.We find, in general, excellent agreements between the theory and the simulations

  3. Improving the Document Development Process: Integrating Relational Data and Statistical Process Control.

    Science.gov (United States)

    Miller, John

    1994-01-01

    Presents an approach to document numbering, document titling, and process measurement which, when used with fundamental techniques of statistical process control, reveals meaningful process-element variation as well as nominal productivity models. (SR)

  4. Thermodynamics, kinetics and process control of nitriding

    DEFF Research Database (Denmark)

    Mittemeijer, Eric J.; Somers, Marcel A. J.

    1999-01-01

    As a prerequisite for predictability of properties obtained by a nitriding treatment of iron-based workpieces, the relation between the process parameters and the composition and structure of the surface layer produced must be known. At present (even) the description of thermodynamic equilibrium...... of pure iron-nitrogen phases has not been achieved fully. It has been shown that taking into account ordering of nitrogen in the epsilon and gamma' iron-nitride phases, leads to an improved understanding of the Fe-N phase diagram. Although thermodynamics indicate the state the system strives for......, the nitriding result is determined largely by the kinetics of the process. The nitriding kinetics have been shown to be characterised by the occurring local near-equilibria and stationary states at surfaces and interfaces, and the diffusion coefficient of nitrogen in the various phases, for which new data have...

  5. Big Data Analytics for Industrial Process Control

    DEFF Research Database (Denmark)

    Khan, Abdul Rauf; Schioler, Henrik; Kulahci, Murat

    2017-01-01

    Today, in modern factories, each step in manufacturing produces a bulk of valuable as well as highly precise information. This provides a great opportunity for understanding the hidden statistical dependencies in the process. Systematic analysis and utilization of advanced analytical methods can ...... lead towards more informed decisions. In this article we discuss some of the challenges related to big data analysis in manufacturing and relevant solutions to some of these challenges....

  6. Human Information Processing and Supervisory Control.

    Science.gov (United States)

    1980-05-01

    errors (that is of the output of the human operator). There is growing evidence (Senders, personal communication; Norman , personal communication...relates to the relative tendency to depend on sensory information or to be more analytic and independent. Norman (personal communication) has referred...decision process model. Ergonomics, 12, 543-557. Senders, J., Elkid, J., Grignetti, M., & Smallwood , R. 1966. An investigation of the visual sampling

  7. Control System Design for Cylindrical Tank Process Using Neural Model Predictive Control Technique

    Directory of Open Access Journals (Sweden)

    M. Sridevi

    2010-10-01

    Full Text Available Chemical manufacturing and process industry requires innovative technologies for process identification. This paper deals with model identification and control of cylindrical process. Model identification of the process was done using ARMAX technique. A neural model predictive controller was designed for the identified model. The performance of the controllers was evaluated using MATLAB software. The performance of NMPC controller was compared with Smith Predictor controller and IMC controller based on rise time, settling time, overshoot and ISE and it was found that the NMPC controller is better suited for this process.

  8. Microeconomics of 300-mm process module control

    Science.gov (United States)

    Monahan, Kevin M.; Chatterjee, Arun K.; Falessi, Georges; Levy, Ady; Stoller, Meryl D.

    2001-08-01

    Simple microeconomic models that directly link metrology, yield, and profitability are rare or non-existent. In this work, we validate and apply such a model. Using a small number of input parameters, we explain current yield management practices in 200 mm factories. The model is then used to extrapolate requirements for 300 mm factories, including the impact of simultaneous technology transitions to 130nm lithography and integrated metrology. To support our conclusions, we use examples relevant to factory-wide photo module control.

  9. Quantum processing by remote quantum control

    Science.gov (United States)

    Qiang, Xiaogang; Zhou, Xiaoqi; Aungskunsiri, Kanin; Cable, Hugo; O'Brien, Jeremy L.

    2017-12-01

    Client-server models enable computations to be hosted remotely on quantum servers. We present a novel protocol for realizing this task, with practical advantages when using technology feasible in the near term. Client tasks are realized as linear combinations of operations implemented by the server, where the linear coefficients are hidden from the server. We report on an experimental demonstration of our protocol using linear optics, which realizes linear combination of two single-qubit operations by a remote single-qubit control. In addition, we explain when our protocol can remain efficient for larger computations, as well as some ways in which privacy can be maintained using our protocol.

  10. Microbial profile and critical control points during processing of 'robo ...

    African Journals Online (AJOL)

    Microbial profile and critical control points during processing of 'robo' snack from ... the relevant critical control points especially in relation to raw materials and ... to the quality of the various raw ingredients used were the roasting using earthen

  11. Advanced Process Control Application and Optimization in Industrial Facilities

    Directory of Open Access Journals (Sweden)

    Howes S.

    2015-01-01

    Full Text Available This paper describes application of the new method and tool for system identification and PID tuning/advanced process control (APC optimization using the new 3G (geometric, gradient, gravity optimization method. It helps to design and implement control schemes directly inside the distributed control system (DCS or programmable logic controller (PLC. Also, the algorithm helps to identify process dynamics in closed-loop mode, optimizes controller parameters, and helps to develop adaptive control and model-based control (MBC. Application of the new 3G algorithm for designing and implementing APC schemes is presented. Optimization of primary and advanced control schemes stabilizes the process and allows the plant to run closer to process, equipment and economic constraints. This increases production rates, minimizes operating costs and improves product quality.

  12. The controlled vitrification/crystallisation process applied

    Directory of Open Access Journals (Sweden)

    Romero, M.

    2000-02-01

    Full Text Available The glass-ceramic process, as well as the usual processing of ceramic and vitreous materials, is being investigated as a promising way for isolation and recycling of both mineral wastes (debris and mineral residues, clearings in public works and inorganic industrial wastes (muds, slags, fly ashes. Synthetic materials with useful properties to be used as building materials have been prepared from inorganic wastes of different type (red muds from zinc hydrometalurgy, fly ashes from power thermal stations, slags and fly ashes from domiciliary incinerators as well as from mixtures of such wastes with other raw materials. The obtained results allow us to conclude that the ceramic and glass-ceramic processes are outlined as an useful alternative to solve the social and environmental problems associated to wastes production.

    El proceso vitrocerámico, así como el procesado habitual de materiales cerámicos y vítreos, está siendo actualmente investigado como una prometedora vía para el aislamiento, inertización e incluso el reciclado de residuos minerales (escombreras y estériles de minas, desmontes de Obras Públicas, etc... e industriales (lodos, fangos, escorias, cenizas, etc.... A partir de residuos inorgánicos de diferente naturaleza (lodos de la hidrometalurgia del zinc, cenizas de centrales térmicas, escorias y cenizas de plantas incineradoras así como de mezclas de los mismos con otras materias primas, se están obteniendo materiales sintéticos con amplias aplicaciones en la Construcción y en Obras Públicas. Los resultados que se están consiguiendo permiten concluir que los procesos cerámico y vitrocerámico se perfilan como una alternativa real y útil para resolver, al menos parcialmente, los problemas sociales y medioambientales asociados a la producción de dichos residuos.

  13. Creys-Malville control room and data processing

    International Nuclear Information System (INIS)

    Decuyper, J.

    1984-01-01

    After a brief definition of the control of a plant, this article presents the Creys-Malville control room: control means display and considerations on ergonomy and specific features in respect of the PWR control room. The Creys-Malville data processing is then rapidly presented with a brief description, the different data treatments and the specificity of the centralised data computer [fr

  14. Plug and Play Process Control of a District Heating System

    DEFF Research Database (Denmark)

    Trangbaek, Klaus; Knudsen, Torben; Skovmose Kallesøe, Carsten

    2009-01-01

    The main idea of plug and play process control is to initialise and reconfigure control systems automatically. In this paper these ideas are applied to a scaled laboratory model of a district heating pressure control system.  First of all this serves as a concrete example of plug and play control...

  15. Memory-type control charts for monitoring the process dispersion

    NARCIS (Netherlands)

    Abbas, N.; Riaz, M.; Does, R.J.M.M.

    2014-01-01

    Control charts have been broadly used for monitoring the process mean and dispersion. Cumulative sum (CUSUM) and exponentially weighted moving average (EWMA) control charts are memory control charts as they utilize the past information in setting up the control structure. This makes CUSUM and

  16. Development of Chemical Process Design and Control for Sustainability

    Science.gov (United States)

    This contribution describes a novel process systems engineering framework that couples advanced control with sustainability evaluation and decision making for the optimization of process operations to minimize environmental impacts associated with products, materials, and energy....

  17. Dosimetry control for radiation processing - basic requirements and standards

    International Nuclear Information System (INIS)

    Ivanova, M.; Tsrunchev, Ts.

    2004-01-01

    A brief review of the basic international codes and standards for dosimetry control for radiation processing (high doses dosimetry), setting up a dosimetry control for radiation processing and metrology control of the dosimetry system is made. The present state of dosimetry control for food processing and the Bulgarian long experience in food irradiation (three irradiation facilities are operational at these moment) are presented. The absence of neither national standard for high doses nor accredited laboratory for calibration and audit of radiation processing dosimetry systems is also discussed

  18. Integrated Intelligent Modeling, Design and Control of Crystal Growth Processes

    National Research Council Canada - National Science Library

    Prasad, V

    2000-01-01

    .... This MURI program took an integrated approach towards modeling, design and control of crystal growth processes and in conjunction with growth and characterization experiments developed much better...

  19. Internal control in the management system of meat processing enterprises

    Directory of Open Access Journals (Sweden)

    Volodymyr Kushnir

    2018-03-01

    Full Text Available The article is described the theoretical basis of internal control and its practical aspects in the work of meat processing enterprises (a case in the meat processing industry in Ukraine. The purpose of the research is to establish the theoretical foundations of the internal control and its improvement in the activity of meat processing plants of various forms of management. It is proposed to use precisely internal control among other names of domestic control. Definition of internal control, its subject and purpose are improved. The subjects and objects of internal control are determined; the principles of its implementation are supplemented. Specific control tasks in meat processing plants according to the needs of this industry are outlined. Specific examples of control subjects are presented and the role of the revision commission is emphasized. The state of internal control in meat processing plants in Ukraine is investigated and it is established that it has a bad condition and unfounded approach to its implementation by managers of meat processing enterprises. To improve the situation we recommend that each meat processing enterprise have in its staff a revision commission or an apposer (auditor. It is established that internal control is more effective in joint-stock companies than in limited liability companies. The necessity of internal control as an important element in the enterprise management system is accented.

  20. High-Level Waste (HLW) Feed Process Control Strategy

    International Nuclear Information System (INIS)

    STAEHR, T.W.

    2000-01-01

    The primary purpose of this document is to describe the overall process control strategy for monitoring and controlling the functions associated with the Phase 1B high-level waste feed delivery. This document provides the basis for process monitoring and control functions and requirements needed throughput the double-shell tank system during Phase 1 high-level waste feed delivery. This document is intended to be used by (1) the developers of the future Process Control Plan and (2) the developers of the monitoring and control system

  1. Safety monitoring in process and control

    International Nuclear Information System (INIS)

    Esparza, V. Jr.; Sebo, D.E.

    1984-01-01

    Safety Functions provide a method of ensuring the safe operation of any large-scale processing plant. Successful implementation of safety functions requires continuous monitoring of safety function values and trends. Because the volume of information handled by a plant operator occassionally can become overwhelming, attention may be diverted from the primary concern of maintaining plant safety. With this in mind EG and G, Idaho developed various methods and techniques for use in a computerized Safety Function Monitoring System and tested the application of these techniques using a simulated nuclear power plant, the Loss-of-Fluid Test Facility (LOFT) at the Idaho National Engineering Laboratory (INEL). This paper presents the methods used in the development of a Safety Function Monitoring System

  2. Control of processes using isotopic diagnostic's technologies

    International Nuclear Information System (INIS)

    Vargas, Celso; Chaverri, Oscar; Chine, Bruno; Conejo, Mario

    2005-01-01

    The Escuela de Ciencias e Ingenieria de los Materiales of the Instituto Tecnologico de Costa Rica, in cooperation with OIEA, develops a project of dowry of capacity oriented to the formation of professionals and equipment for the use of two important technologies of isotopic diagnostic. The first of them is the technology of tracers that operates the unique properties that present different radioactive isotopics like open sources. The second one well known as scanning or profile gamma, uses sealed source, of the some nature that the previous ones, to obtain profiles in different processes and thus to determine its internal condition and operation. The objective of this article is to present both technologies, its benefits and to promote the use in the country [es

  3. Case Studies in Modelling, Control in Food Processes.

    Science.gov (United States)

    Glassey, J; Barone, A; Montague, G A; Sabou, V

    This chapter discusses the importance of modelling and control in increasing food process efficiency and ensuring product quality. Various approaches to both modelling and control in food processing are set in the context of the specific challenges in this industrial sector and latest developments in each area are discussed. Three industrial case studies are used to demonstrate the benefits of advanced measurement, modelling and control in food processes. The first case study illustrates the use of knowledge elicitation from expert operators in the process for the manufacture of potato chips (French fries) and the consequent improvements in process control to increase the consistency of the resulting product. The second case study highlights the economic benefits of tighter control of an important process parameter, moisture content, in potato crisp (chips) manufacture. The final case study describes the use of NIR spectroscopy in ensuring effective mixing of dry multicomponent mixtures and pastes. Practical implementation tips and infrastructure requirements are also discussed.

  4. Represenation of process state, structure and control

    International Nuclear Information System (INIS)

    Goodstein, L.P.; Rasmussen, J.

    1987-04-01

    Supervisory control is essentially a decision-making activity where, among other things, the dm has to maneuver within a complex problem space which reflects key dimensions and attributes of the object system (power plant ...). Of considerable importance therefore is the representation for the dm of this problem space comprising at the one end the target demands, goals, constraints and, at the other, the resources available for meeting the assigned goals - and all of this in pace with the dynamic event-driven environment which characterizes the types of systems of interest. Previous work has identified the advantages of utilizing the two-dimensional means-ends/part-whole space as a basic ingredient in a system representation. This paper associates more detailed representational requirements at the various levels of the means-ends axis with the activities of state identification and diagnosis. In addition, some examples of display formats which attempt to incorporate the outlined representational principles within the context of a PWR plant are discussed. (authors)

  5. Positive affect improves working memory: implications for controlled cognitive processing.

    Science.gov (United States)

    Yang, Hwajin; Yang, Sujin; Isen, Alice M

    2013-01-01

    This study examined the effects of positive affect on working memory (WM) and short-term memory (STM). Given that WM involves both storage and controlled processing and that STM primarily involves storage processing, we hypothesised that if positive affect facilitates controlled processing, it should improve WM more than STM. The results demonstrated that positive affect, compared with neutral affect, significantly enhanced WM, as measured by the operation span task. The influence of positive affect on STM, however, was weaker. These results suggest that positive affect enhances WM, a task that involves controlled processing, not just storage processing. Additional analyses of recall and processing times and accuracy further suggest that improved WM under positive affect is not attributable to motivational differences, but results instead from improved controlled cognitive processing.

  6. PROCESS CONTROL STORAGE SMOKELESS TOBACCO (SNUS

    Directory of Open Access Journals (Sweden)

    T. A. Don

    2015-01-01

    Full Text Available An important aim during smokeless tobacco product (snus storage is keeping its quantitative and qualitative properties. Another significant factor is its biological safety due to its oral way of consumption. The main factors affecting quality during storage are: temperature, relative humidity, air composition, ventilation rate, surrounding products and package. Snus consumption is growing, but till present it was slightly spread in Russia and as a result little studied, thus studying process of its storage is an actual question. The aims of the research were: studying terms and methods of storage that provide constant product’s quality. Samples with different composition were taken for carrying research. Quality evaluation was carried according to standard for tobacco branch methods and to new methods developed in the laboratory of technologies for manufacturing tobacco products. Products were stored at lowered and room temperatures. It is discovered that snus with herbs added can be stored at room temperature and humidity not more than three months. Snus with coffee added can be stored not more than four months without changing its taste. Adding honey to this product slightly decreases its score by sensory evaluation, but remains it possible for consumption. Sensory testing score after storage at lowered temperature has remained unchanged. Optimal terms of storage were found. For room temperature they are: temperature 17-25 °C, relative air humidity 65 – 75 %, for lowered temperature – 5 °C and relative air humidity 50 -60 %. It was discovered that composition of snus and environmental parameters affect its storage duration.

  7. SPS/LEP beam transfer equipment control using industrial automation components

    International Nuclear Information System (INIS)

    Aimar, A.; Berard, G.; Bretin, J.L.; Carlier, E.; Dieperink, J.H.; Laffin, M.; Mertens, V.; Verhagen, H.

    1992-01-01

    Several control systems for SPS and LEP beam transfer equipment have to be commissioned in the near future. Tools for fast software development, easy maintenance and modifications, compliance with industrial standards, and independence of specific suppliers are considered to be essential. A large fraction of the systems can be realized using off-the-shelf industrial automation components like industrial I/O systems, programmable logic controllers, or diskless PCs. Specific electronics built up in G-64 can be integrated. Diskless systems running UNIX and X Windows are foreseen as process controllers and local access media. (author)

  8. 10 CFR 72.158 - Control of special processes.

    Science.gov (United States)

    2010-01-01

    ... 10 Energy 2 2010-01-01 2010-01-01 false Control of special processes. 72.158 Section 72.158 Energy... NUCLEAR FUEL, HIGH-LEVEL RADIOACTIVE WASTE, AND REACTOR-RELATED GREATER THAN CLASS C WASTE Quality Assurance § 72.158 Control of special processes. The licensee, applicant for a license, certificate holder...

  9. Distributed automatic control of technological processes in conditions of weightlessness

    Science.gov (United States)

    Kukhtenko, A. I.; Merkulov, V. I.; Samoylenko, Y. I.; Ladikov-Royev, Y. P.

    1986-01-01

    Some problems associated with the automatic control of liquid metal and plasma systems under conditions of weightlessness are examined, with particular reference to the problem of stability of liquid equilibrium configurations. The theoretical fundamentals of automatic control of processes in electrically conducting continuous media are outlined, and means of using electromagnetic fields for simulating technological processes in a space environment are discussed.

  10. Using Paper Helicopters to Teach Statistical Process Control

    Science.gov (United States)

    Johnson, Danny J.

    2011-01-01

    This hands-on project uses a paper helicopter to teach students how to distinguish between common and special causes of variability when developing and using statistical process control charts. It allows the student to experience a process that is out-of-control due to imprecise or incomplete product design specifications and to discover how the…

  11. Using Statistical Process Control to Enhance Student Progression

    Science.gov (United States)

    Hanna, Mark D.; Raichura, Nilesh; Bernardes, Ednilson

    2012-01-01

    Public interest in educational outcomes has markedly increased in the most recent decade; however, quality management and statistical process control have not deeply penetrated the management of academic institutions. This paper presents results of an attempt to use Statistical Process Control (SPC) to identify a key impediment to continuous…

  12. Process theory for supervisory control of stochastic systems with data

    NARCIS (Netherlands)

    Markovski, J.

    2012-01-01

    We propose a process theory for supervisory control of stochastic nondeterministic plants with data-based observations. The Markovian process theory with data relies on the notion of Markovian partial bisimulation to capture controllability of stochastic nondeterministic systems. It presents a

  13. Manufacturing Squares: An Integrative Statistical Process Control Exercise

    Science.gov (United States)

    Coy, Steven P.

    2016-01-01

    In the exercise, students in a junior-level operations management class are asked to manufacture a simple product. Given product specifications, they must design a production process, create roles and design jobs for each team member, and develop a statistical process control plan that efficiently and effectively controls quality during…

  14. State Space Reduction of Linear Processes using Control Flow Reconstruction

    NARCIS (Netherlands)

    van de Pol, Jan Cornelis; Timmer, Mark

    2009-01-01

    We present a new method for fighting the state space explosion of process algebraic specifications, by performing static analysis on an intermediate format: linear process equations (LPEs). Our method consists of two steps: (1) we reconstruct the LPE's control flow, detecting control flow parameters

  15. State Space Reduction of Linear Processes Using Control Flow Reconstruction

    NARCIS (Netherlands)

    van de Pol, Jan Cornelis; Timmer, Mark; Liu, Zhiming; Ravn, Anders P.

    2009-01-01

    We present a new method for fighting the state space explosion of process algebraic specifications, by performing static analysis on an intermediate format: linear process equations (LPEs). Our method consists of two steps: (1) we reconstruct the LPE's control flow, detecting control flow parameters

  16. Nitrate control strategies in an activated sludge wastewater treatment process

    Energy Technology Data Exchange (ETDEWEB)

    Shen, Wenhao; Tao, Erpan; Chen, Xiaoquan; Liu, Dawei [South China University of Technology, Guangzhou (China); Liu, Hongbin [Kyung Hee University, Yongin (Korea, Republic of)

    2014-03-15

    We studied nitrate control strategies in an activated sludge wastewater treatment process (WWTP) based on the activated sludge model. Two control strategies, back propagation for proportional-integral-derivative (BP-PID) and adaptive-network based fuzzy inference systems (ANFIS), are applied in the WWTP. The simulation results show that the simple local constant setpoint control has poor control effects on the nitrate concentration control. However, the ANFIS (4*1) controller, which considers not only the local constant setpoint control of the nitrate concentration, but also three important indices in the effluent--ammonia concentration, total suspended sludge concentration and total nitrogen concentration--demonstrates good control performance. The results also prove that ANFIS (4*1) controller has better control performance than that of the controllers PI, BP-PID and ANFIS (2*1), and that the ANFIS (4*1) controller is effective in improving the effluent quality and maintaining the stability of the effluent quality.

  17. Development and upgrade of new real time processor in JT-60 data processing system

    International Nuclear Information System (INIS)

    Sakata, Shinya; Koiwa, Motonao; Matsuda, Toshiaki; Aoyagi, Tetsuo

    2000-07-01

    At the beginning of JT-60 experiments, the real time processor (RTP) in the data processing system was mainly constructed by PANAFACOM U-1500. As the computer became superannuated, however, it gradually became difficult to maintain both hardware and software. A performance of a recent UNIX workstation has been remarkably progressed. The UNIX workstation has a large flexibility for user application programs, an easiness for maintenance of the hardware and an ability of expansion to peripheral devices. Therefore, the RTP system is newly reconstructed by using the UNIX workstation. This report describes the overview, the basic design and the recent upgrade on the RTP in the data processing system. (author)

  18. Process control analysis of IMRT QA: implications for clinical trials

    International Nuclear Information System (INIS)

    Pawlicki, Todd; Rice, Roger K; Yoo, Sua; Court, Laurence E; McMillan, Sharon K; Russell, J Donald; Pacyniak, John M; Woo, Milton K; Basran, Parminder S; Boyer, Arthur L; Bonilla, Claribel

    2008-01-01

    The purpose of this study is two-fold: first is to investigate the process of IMRT QA using control charts and second is to compare control chart limits to limits calculated using the standard deviation (σ). Head and neck and prostate IMRT QA cases from seven institutions in both academic and community settings are considered. The percent difference between the point dose measurement in phantom and the corresponding result from the treatment planning system (TPS) is used for analysis. The average of the percent difference calculations defines the accuracy of the process and is called the process target. This represents the degree to which the process meets the clinical goal of 0% difference between the measurements and TPS. IMRT QA process ability defines the ability of the process to meet clinical specifications (e.g. 5% difference between the measurement and TPS). The process ability is defined in two ways: (1) the half-width of the control chart limits, and (2) the half-width of ±3σ limits. Process performance is characterized as being in one of four possible states that describes the stability of the process and its ability to meet clinical specifications. For the head and neck cases, the average process target across institutions was 0.3% (range: -1.5% to 2.9%). The average process ability using control chart limits was 7.2% (range: 5.3% to 9.8%) compared to 6.7% (range: 5.3% to 8.2%) using standard deviation limits. For the prostate cases, the average process target across the institutions was 0.2% (range: -1.8% to 1.4%). The average process ability using control chart limits was 4.4% (range: 1.3% to 9.4%) compared to 5.3% (range: 2.3% to 9.8%) using standard deviation limits. Using the standard deviation to characterize IMRT QA process performance resulted in processes being preferentially placed in one of the four states. This is in contrast to using control charts for process characterization where the IMRT QA processes were spread over three of the

  19. Process Control System Cyber Security Standards - An Overview

    Energy Technology Data Exchange (ETDEWEB)

    Robert P. Evans

    2006-05-01

    The use of cyber security standards can greatly assist in the protection of process control systems by providing guidelines and requirements for the implementation of computer-controlled systems. These standards are most effective when the engineers and operators, using the standards, understand what each standard addresses. This paper provides an overview of several standards that deal with the cyber security of process measurements and control systems.

  20. Iterative Controller Tuning for Process with Fold Bifurcations

    DEFF Research Database (Denmark)

    Huusom, Jakob Kjøbsted; Poulsen, Niels Kjølstad; Jørgensen, Sten Bay

    2007-01-01

    Processes involving fold bifurcation are notoriously difficult to control in the vicinity of the fold where most often optimal productivity is achieved . In cases with limited process insight a model based control synthesis is not possible. This paper uses a data driven approach with an improved...... version of iterative feedback tuning to optimizing a closed loop performance criterion, as a systematic tool for tuning process with fold bifurcations....

  1. Control of instability in nitric acid evaporators for plutonium processing

    International Nuclear Information System (INIS)

    1998-03-01

    Improved control of the nitric acid process evaporators requires the detection of spontaneously unstable operating conditions. This process reduces the volume of contaminated liquid by evaporating nitric acid and concentrating salt residues. If a instability is identified quickly, prompt response can avert distillate contamination. An algorithm applied to the runtime data was evaluated to detect this situation. A snapshot of data from a histogram in the old process control software was captured during the unstable conditions and modeled

  2. Applying Statistical Process Quality Control Methodology to Educational Settings.

    Science.gov (United States)

    Blumberg, Carol Joyce

    A subset of Statistical Process Control (SPC) methodology known as Control Charting is introduced. SPC methodology is a collection of graphical and inferential statistics techniques used to study the progress of phenomena over time. The types of control charts covered are the null X (mean), R (Range), X (individual observations), MR (moving…

  3. Diagnostic system for process control at NPP Dukovany load follow

    International Nuclear Information System (INIS)

    Rubek, J.; Petruzela, I.

    1998-01-01

    The NPP Dukovany is being operated in the frequency control since 1996. In last year a project for the plant load follow has been developed. One part of the project is to install a diagnostic system for process control. At present the main control loops of the plant control system are regular tested after unit refuelling only. The functionality and control system parameter adjusting is tested by certificated procedures. This state is unsuitable in view of the plan load follow operation. The relevant operational modes are based on minimisation of influence on plant component life time and on achievement of planned unit parameters. Therefore it is necessary to provide testing of main control system parts in shorter time period. Mainly at time when the unit is really in load follow operation. The paper describes the diagnostic system for process control which will be at NPP Dukovany implemented. The principal of the system will be evaluation of real and expected changes of technological variables. The system utilises thermohydraulic relation among main technological variables and relation among controlled and manipulated variables. Outputs of the system will be used to operational staff support at the plant operation. It enables: determination of control system state, estimation and check of future control system state, early indication of the deviation of process from normal conditions, check of efficiency of operational staff intervention into plant control. The system gives the plant operator new information for the plant process control. Simultaneously the coupling of new system outputs on existing signalisation is solved. (author)

  4. Process-based quality for thermal spray via feedback control

    Science.gov (United States)

    Dykhuizen, R. C.; Neiser, R. A.

    2006-09-01

    Quality control of a thermal spray system manufacturing process is difficult due to the many input variables that need to be controlled. Great care must be taken to ensure that the process remains constant to obtain a consistent quality of the parts. Control is greatly complicated by the fact that measurement of particle velocities and temperatures is a noisy stochastic process. This article illustrates the application of quality control concepts to a wire flame spray process. A central feature of the real-time control system is an automatic feedback control scheme that provides fine adjustments to ensure that uncontrolled variations are accommodated. It is shown how the control vectors can be constructed from simple process maps to independently control particle velocity and temperature. This control scheme is shown to perform well in a real production environment. We also demonstrate that slight variations in the feed wire curvature can greatly influence the process. Finally, the geometry of the spray system and sensor must remain constant for the best reproducibility.

  5. Neural Correlates of Automatic and Controlled Auditory Processing in Schizophrenia

    Science.gov (United States)

    Morey, Rajendra A.; Mitchell, Teresa V.; Inan, Seniha; Lieberman, Jeffrey A.; Belger, Aysenil

    2009-01-01

    Individuals with schizophrenia demonstrate impairments in selective attention and sensory processing. The authors assessed differences in brain function between 26 participants with schizophrenia and 17 comparison subjects engaged in automatic (unattended) and controlled (attended) auditory information processing using event-related functional MRI. Lower regional neural activation during automatic auditory processing in the schizophrenia group was not confined to just the temporal lobe, but also extended to prefrontal regions. Controlled auditory processing was associated with a distributed frontotemporal and subcortical dysfunction. Differences in activation between these two modes of auditory information processing were more pronounced in the comparison group than in the patient group. PMID:19196926

  6. Materials of the Regional Training Course on Validation and Process Control for Electron Beam Radiation Processing

    International Nuclear Information System (INIS)

    Kaluska, I.; Gluszewski, W.

    2007-01-01

    Irradiation with electron beams is used in the polymer industry, food, pharmaceutical and medical device industries for sterilization of surfaces. About 20 lectures presented during the Course were devoted to all aspects of control and validation of low energy electron beam processes. They should help the product manufacturers better understand the application of the ANSI/AAMI/ISO 11137 norm, which defines the requirements and standard practices for validation of the irradiation process and the process controls required during routine processing

  7. A case study: application of statistical process control tool for determining process capability and sigma level.

    Science.gov (United States)

    Chopra, Vikram; Bairagi, Mukesh; Trivedi, P; Nagar, Mona

    2012-01-01

    Statistical process control is the application of statistical methods to the measurement and analysis of variation process. Various regulatory authorities such as Validation Guidance for Industry (2011), International Conference on Harmonisation ICH Q10 (2009), the Health Canada guidelines (2009), Health Science Authority, Singapore: Guidance for Product Quality Review (2008), and International Organization for Standardization ISO-9000:2005 provide regulatory support for the application of statistical process control for better process control and understanding. In this study risk assessments, normal probability distributions, control charts, and capability charts are employed for selection of critical quality attributes, determination of normal probability distribution, statistical stability, and capability of production processes, respectively. The objective of this study is to determine tablet production process quality in the form of sigma process capability. By interpreting data and graph trends, forecasting of critical quality attributes, sigma process capability, and stability of process were studied. The overall study contributes to an assessment of process at the sigma level with respect to out-of-specification attributes produced. Finally, the study will point to an area where the application of quality improvement and quality risk assessment principles for achievement of six sigma-capable processes is possible. Statistical process control is the most advantageous tool for determination of the quality of any production process. This tool is new for the pharmaceutical tablet production process. In the case of pharmaceutical tablet production processes, the quality control parameters act as quality assessment parameters. Application of risk assessment provides selection of critical quality attributes among quality control parameters. Sequential application of normality distributions, control charts, and capability analyses provides a valid statistical

  8. Modeling and Control of Multivariable Process Using Intelligent Techniques

    Directory of Open Access Journals (Sweden)

    Subathra Balasubramanian

    2010-10-01

    Full Text Available For nonlinear dynamic systems, the first principles based modeling and control is difficult to implement. In this study, a fuzzy controller and recurrent fuzzy controller are developed for MIMO process. Fuzzy logic controller is a model free controller designed based on the knowledge about the process. In fuzzy controller there are two types of rule-based fuzzy models are available: one the linguistic (Mamdani model and the other is Takagi–Sugeno model. Of these two, Takagi-Sugeno model (TS has attracted most attention. The fuzzy controller application is limited to static processes due to their feedforward structure. But, most of the real-time processes are dynamic and they require the history of input/output data. In order to store the past values a memory unit is needed, which is introduced by the recurrent structure. The proposed recurrent fuzzy structure is used to develop a controller for the two tank heating process. Both controllers are designed and implemented in a real time environment and their performance is compared.

  9. Gemstone Grinding Process Improvement by using Impedance Force Control

    Directory of Open Access Journals (Sweden)

    Hamprommarat Chumpol

    2015-01-01

    Full Text Available Chula Automatic Faceting Machine has been developed by The Advance Manufacturing Research Lab, Chulalongkorn University to support Thailand Gems-Industry. The machine has high precision motion control by using position and force control. A contact stiffness model is used to estimate grinding force. Although polished gems from the Faceting Machine have uniform size and acceptable shape, the force of the grinding and polishing process cannot be maintain constant and has some fluctuation due to indirect force control. Therefor this research work propose a new controller for this process based on an impedance direct force control to improve the gemstone grinding performance during polishing process. The grinding force can be measured through motor current. The results show that the polished gems by using impedance direct force control can maintain uniform size as well as good shape and high quality surface.

  10. Expert system and process optimization techniques for real-time monitoring and control of plasma processes

    Science.gov (United States)

    Cheng, Jie; Qian, Zhaogang; Irani, Keki B.; Etemad, Hossein; Elta, Michael E.

    1991-03-01

    To meet the ever-increasing demand of the rapidly-growing semiconductor manufacturing industry it is critical to have a comprehensive methodology integrating techniques for process optimization real-time monitoring and adaptive process control. To this end we have accomplished an integrated knowledge-based approach combining latest expert system technology machine learning method and traditional statistical process control (SPC) techniques. This knowledge-based approach is advantageous in that it makes it possible for the task of process optimization and adaptive control to be performed consistently and predictably. Furthermore this approach can be used to construct high-level and qualitative description of processes and thus make the process behavior easy to monitor predict and control. Two software packages RIST (Rule Induction and Statistical Testing) and KARSM (Knowledge Acquisition from Response Surface Methodology) have been developed and incorporated with two commercially available packages G2 (real-time expert system) and ULTRAMAX (a tool for sequential process optimization).

  11. The CANDU 9 distributed control system design process

    International Nuclear Information System (INIS)

    Harber, J.E.; Kattan, M.K.; Macbeth, M.J.

    1997-01-01

    Canadian designed CANDU pressurized heavy water nuclear reactors have been world leaders in electrical power generation. The CANDU 9 project is AECL's next reactor design. Plant control for the CANDU 9 station design is performed by a distributed control system (DCS) as compared to centralized control computers, analog control devices and relay logic used in previous CANDU designs. The selection of a DCS as the platform to perform the process control functions and most of the data acquisition of the plant, is consistent with the evolutionary nature of the CANDU technology. The control strategies for the DCS control programs are based on previous CANDU designs but are implemented on a new hardware platform taking advantage of advances in computer technology. This paper describes the design process for developing the CANDU 9 DCS. Various design activities, prototyping and analyses have been undertaken in order to ensure a safe, functional, and cost-effective design. (author)

  12. Interfacing industrial process control systems to LEP/LHC

    International Nuclear Information System (INIS)

    Rabany, M.

    1992-01-01

    Modern industrial process control systems have developed to meet the needs of industry to increase the production while decreasing the costs. Although particle accelerators designers have pioneered in control systems during the seventies, it has now become possible to them to profit of industrial solutions in substitution of, or in complement with the more traditional home made ones. Adapting and integrating such industrial systems to the accelerator control area will certainly benefit to the field in terms of finance, human resources and technical facilities offered off-the-shelf by the widely experienced industrial controls community; however this cannot be done without slightly affecting the overall accelerator control architecture. The paper briefly describes the industrial controls arena and takes example on an industrial process control system recently installed at CERN to discuss in detail the related choices and issues. (author)

  13. Software control and system configuration management - A process that works

    Science.gov (United States)

    Petersen, K. L.; Flores, C., Jr.

    1983-01-01

    A comprehensive software control and system configuration management process for flight-crucial digital control systems of advanced aircraft has been developed and refined to insure efficient flight system development and safe flight operations. Because of the highly complex interactions among the hardware, software, and system elements of state-of-the-art digital flight control system designs, a systems-wide approach to configuration control and management has been used. Specific procedures are implemented to govern discrepancy reporting and reconciliation, software and hardware change control, systems verification and validation testing, and formal documentation requirements. An active and knowledgeable configuration control board reviews and approves all flight system configuration modifications and revalidation tests. This flexible process has proved effective during the development and flight testing of several research aircraft and remotely piloted research vehicles with digital flight control systems that ranged from relatively simple to highly complex, integrated mechanizations.

  14. Process control of an HTGR fuel reprocessing cold pilot plant

    International Nuclear Information System (INIS)

    Rode, J.S.

    1976-10-01

    Development of engineering-scale systems for a large-scale HTGR fuel reprocessing demonstration facility is currently underway in a cold pilot plant. These systems include two fluidized-bed burners, which remove the graphite (carbon) matrix from the crushed HTGR fuel by high temperature (900 0 C) oxidation. The burners are controlled by a digital process controller with an all analog input/output interface which has been in use since March, 1976. The advantages of such a control system to a pilot plant operation can be summarized as follows: (1) Control loop functions and configurations can be changed easily; (2) control constants, alarm limits, output limits, and scaling constants can be changed easily; (3) calculation of data and/or interface with a computerized information retrieval system during operation are available; (4) diagnosis of process control problems is facilitated; and (5) control panel/room space is saved

  15. New Process Controls for the Hera Cryogenic Plant

    Science.gov (United States)

    Böckmann, T.; Clausen, M.; Gerke, Chr.; Prüß, K.; Schoeneburg, B.; Urbschat, P.

    2010-04-01

    The cryogenic plant built for the HERA accelerator at DESY in Hamburg (Germany) is now in operation for more than two decades. The commercial process control system for the cryogenic plant is in operation for the same time period. Ever since the operator stations, the control network and the CPU boards in the process controllers went through several upgrade stages. Only the centralized Input/Output system was kept unchanged. Many components have been running beyond the expected lifetime. The control system for one at the three parts of the cryogenic plant has been replaced recently by a distributed I/O system. The I/O nodes are connected to several Profibus-DP field busses. Profibus provides the infrastructure to attach intelligent sensors and actuators directly to the process controllers which run the open source process control software EPICS. This paper describes the modification process on all levels from cabling through I/O configuration, the process control software up to the operator displays.

  16. Process control measurements in the SRP fuel separations plants

    International Nuclear Information System (INIS)

    McKibben, J.M.; Pickett, C.E.; Dickert, H.D.

    1982-02-01

    Programs were started to develop new in-line and at-line analytical techniques. Among the more promising techniques being investigated are: (1) an in-line instrument to analyze for percent tributyl phosphate in process solvent, (2) remote laser optrode techniques (using lazer light transmitted to and from the sample cell via light pipes) for a variety of possible analyses, and (3) sonic techniques for concentration analyses in two component systems. A subcommittee was also formed to investigate state-of-the-technology for process control. The final recommendation was to use a distributed control approach to upgrade the process control sytem. The system selected should be modular, easy to expand, and simple to change control strategies. A distributed system using microprocessorbased controllers would allow installation of the control intelligence near the process, thereby simplifying field wiring. Process information collected and stored in the controllers will be transmitted to operating consoles, via a data highway, for process management and display. The overall program has a number of distinct benefits. There are a number of cost savings that will be realized. Excellent annual return on investment - up to 110% - has been predicted for several of the projects in this program that are already funded. In addition, many of the instrument modifications will improve safety performance and production throughput in the specific ways shown

  17. A fast PID controller Design for Modern PLC for Process Control Application

    International Nuclear Information System (INIS)

    Mirza, A.; Nafis, A.; Anees, R.M.; Idris, S.

    2004-01-01

    PID is the most widely used control scheme in the process industry. Pill controllers are utilized for the control of such varied parameters as pressure, flow, temperature, etc. One characteristic of these parameters is that they posses slow dynamics. Most of the available digital controllers can manipulate only a single parameter- multiple controllers are required for control of more than one parameter. The Fast PID Controller for Modem PLC (Programmable Logic Controller) developed by the authors, provides control of several parameters at a time (through a single Pill control element), enhanced programmability including variable sampling period, parameter monitoring and data storage, which may be easily implemented in a PLC. (author)

  18. On-line optimal control improves gas processing

    International Nuclear Information System (INIS)

    Berkowitz, P.N.; Papadopoulos, M.N.

    1992-01-01

    This paper reports that the authors' companies jointly funded the first phase of a gas processing liquids optimization project that has the specific purposes to: Improve the return of processing natural gas liquids, Develop sets of control algorithms, Make available a low-cost solution suitable for small to medium-sized gas processing plants, Test and demonstrate the feasibility of line control. The ARCO Willard CO 2 gas recovery processing plant was chosen as the initial test site to demonstrate the application of multivariable on-line optimal control. One objective of this project is to support an R ampersand D effort to provide a standardized solution to the various types of gas processing plants in the U.S. Processes involved in these gas plants include cryogenic separations, demethanization, lean oil absorption, fractionation and gas treating. Next, the proposed solutions had to be simple yet comprehensive enough to allow an operator to maintain product specifications while operating over a wide range of gas input flow and composition. This had to be a supervisors system that remained on-line more than 95% of the time, and achieved reduced plant operating variability and improved variable cost control. It took more than a year to study various gas processes and to develop a control approach before a real application was finally exercised. An initial process for C 2 and CO 2 recoveries was chosen

  19. Energy Efficient Pump Control for an Offshore Oil Processing System

    DEFF Research Database (Denmark)

    Yang, Zhenyu; Soleiman, Kian; Løhndorf, Bo

    2012-01-01

    The energy efficient control of a pump system for an offshore oil processing system is investigated. The seawater is lifted up by a pump system which consists of three identical centrifugal pumps in parallel, and the lifted seawater is used to cool down the crude oil flowing out of a threephase...... separator on one of the Danish north-sea platform. A hierarchical pump-speed control strategy is developed for the considered system by minimizing the pump power consumption subject to keeping a satisfactory system performance. The proposed control strategy consists of online estimation of some system...... operating parameters, optimization of pump configurations, and a real-time feedback control. Comparing with the current control strategy at the considered system, where the pump system is on/off controlled, and the seawater flows are controlled by a number of control valves, the proposed control strategy...

  20. Development, validation and routine control of a radiation process

    International Nuclear Information System (INIS)

    Kishor Mehta

    2010-01-01

    Today, radiation is used in industrial processing for variety of applications; from low doses for blood irradiation to very high doses for materials modification and even higher for gemstone colour enhancement. At present, radiation is mainly provided by either radionuclides or machine sources; cobalt-60 is the most predominant radionuclide in use. Currently, there are several hundred irradiation facilities worldwide. Similar to other industries, quality management systems can assist radiation processing facilities in enhancing customer satisfaction and maintaining and improving product quality. To help fulfill quality management requirements, several national and international organizations have developed various standards related to radiation processing. They all have requirements and guidelines for development, validation and routine control of the radiation process. For radiation processing, these three phases involve the following activities. Development phase includes selecting the type of radiation source, irradiation facility and the dose required for the process. Validation phase includes conducting activities that give assurance that the process will be successful. Routine control then involves activities that provide evidence that the process has been successfully realized. These standards require documentary evidence that process validation and process control have been followed. Dosimetry information gathered during these processes provides this evidence. (authors)

  1. Process control monitoring systems, industrial plants, and process control monitoring methods

    Science.gov (United States)

    Skorpik, James R [Kennewick, WA; Gosselin, Stephen R [Richland, WA; Harris, Joe C [Kennewick, WA

    2010-09-07

    A system comprises a valve; a plurality of RFID sensor assemblies coupled to the valve to monitor a plurality of parameters associated with the valve; a control tag configured to wirelessly communicate with the respective tags that are coupled to the valve, the control tag being further configured to communicate with an RF reader; and an RF reader configured to selectively communicate with the control tag, the reader including an RF receiver. Other systems and methods are also provided.

  2. Optimal control of switched systems arising in fermentation processes

    CERN Document Server

    Liu, Chongyang

    2014-01-01

    The book presents, in a systematic manner, the optimal controls under different mathematical models in fermentation processes. Variant mathematical models – i.e., those for multistage systems; switched autonomous systems; time-dependent and state-dependent switched systems; multistage time-delay systems and switched time-delay systems – for fed-batch fermentation processes are proposed and the theories and algorithms of their optimal control problems are studied and discussed. By putting forward novel methods and innovative tools, the book provides a state-of-the-art and comprehensive systematic treatment of optimal control problems arising in fermentation processes. It not only develops nonlinear dynamical system, optimal control theory and optimization algorithms, but can also help to increase productivity and provide valuable reference material on commercial fermentation processes.

  3. Using Statistical Process Control Methods to Classify Pilot Mental Workloads

    National Research Council Canada - National Science Library

    Kudo, Terence

    2001-01-01

    .... These include cardiac, ocular, respiratory, and brain activity measures. The focus of this effort is to apply statistical process control methodology on different psychophysiological features in an attempt to classify pilot mental workload...

  4. The Use of Statistical Methods in Dimensional Process Control

    National Research Council Canada - National Science Library

    Krajcsik, Stephen

    1985-01-01

    ... erection. To achieve this high degree of unit accuracy, we have begun a pilot dimensional control program that has set the guidelines for systematically monitoring each stage of the production process prior to erection...

  5. Two-process approach to electron beam welding control

    International Nuclear Information System (INIS)

    Lastovirya, V.N.

    1987-01-01

    The analysis and synthesis of multi-dimensional welding control systems, which require the usage of computers, should be conducted within the temporal range. From the general control theory point two approaches - one-process and two-process - are possible to electron beam welding. In case of two-process approach, subprocesses of heat source formation and direct metal melting are separated. Two-process approach leads to two-profile control system and provides the complete controlability of electron beam welding within the frameworks of systems with concentrated, as well as, with distributed parameters. Approach choice for the given problem solution is determined, first of all, by stability degree of heat source during welding

  6. A reconfigurable hybrid supervisory system for process control

    International Nuclear Information System (INIS)

    Garcia, H.E.; Ray, A.; Edwards, R.M.

    1994-01-01

    This paper presents a reconfigurable approach to decision and control systems for complex dynamic processes. The proposed supervisory control system is a reconfigurable hybrid architecture structured into three functional levels of hierarchy, namely, execution, supervision, and coordination. While the bottom execution level is constituted by either reconfigurable continuously varying or discrete event systems, the top two levels are necessarily governed by reconfigurable sets of discrete event decision and control systems. Based on the process status, the set of active control and supervisory algorithm is chosen. The reconfigurable hybrid system is briefly described along with a discussion on its implementation at the Experimental Breeder Reactor II of Argonne National Laboratory. A process control application of this hybrid system is presented and evaluated in an in-plant experiment

  7. A reconfigurable hybrid supervisory system for process control

    International Nuclear Information System (INIS)

    Garcia, H.E.; Ray, A.; Edwards, R.M.

    1994-01-01

    This paper presents a reconfigurable approach to decision and control systems for complex dynamic processes. The proposed supervisory control system is a reconfigurable hybrid architecture structured into three functional levels of hierarchy, namely, execution, supervision, and coordination. While, the bottom execution level is constituted by either reconfigurable continuously varying or discrete event systems, the top two levels are necessarily governed by reconfigurable sets of discrete event decision and control systems. Based on the process status, the set of active control and supervisory algorithm is chosen. The reconfigurable hybrid system is briefly described along with a discussion on its implementation at the Experimental Breeder Reactor 2 of Argonne National Laboratory. A process control application of this hybrid system is presented and evaluated in an in-plant experiment

  8. Instrumentation and control for fossil-energy processes

    Energy Technology Data Exchange (ETDEWEB)

    1982-09-01

    The 1982 symposium on instrumentation and control for fossil energy processes was held June 7 through 9, 1982, at Adam's Mark Hotel, Houston, Texas. It was sponsored by the US Department of Energy, Office of Fossil Energy; Argonne National Laboratory; and the Society for Control and Instrumentation of Energy Processes. Fifty-two papers have been entered individually into EDB and ERA; eleven papers had been entered previously from other sources. (LTN)

  9. Radiation process control, study and acceptance of dosimetric methods

    International Nuclear Information System (INIS)

    Radak, B.B.

    1984-01-01

    The methods of primary dosimetric standardization and the calibration of dosimetric monitors suitable for radiation process control were outlined in the form of a logical pattern in which they are in current use on industrial scale in Yugoslavia. The reliability of the process control of industrial sterilization of medical supplies for the last four years was discussed. The preparatory works for the intermittent use of electron beams in cable industry were described. (author)

  10. Intelligent Transportation Control based on Proactive Complex Event Processing

    OpenAIRE

    Wang Yongheng; Geng Shaofeng; Li Qian

    2016-01-01

    Complex Event Processing (CEP) has become the key part of Internet of Things (IoT). Proactive CEP can predict future system states and execute some actions to avoid unwanted states which brings new hope to intelligent transportation control. In this paper, we propose a proactive CEP architecture and method for intelligent transportation control. Based on basic CEP technology and predictive analytic technology, a networked distributed Markov decision processes model with predicting states is p...

  11. Process control and optimization with simple interval calculation method

    DEFF Research Database (Denmark)

    Pomerantsev, A.; Rodionova, O.; Høskuldsson, Agnar

    2006-01-01

    for the quality improvement in the course of production. The latter is an active quality optimization, which takes into account the actual history of the process. The advocate approach is allied to the conventional method of multivariate statistical process control (MSPC) as it also employs the historical process......Methods of process control and optimization are presented and illustrated with a real world example. The optimization methods are based on the PLS block modeling as well as on the simple interval calculation methods of interval prediction and object status classification. It is proposed to employ...... the series of expanding PLS/SIC models in order to support the on-line process improvements. This method helps to predict the effect of planned actions on the product quality and thus enables passive quality control. We have also considered an optimization approach that proposes the correcting actions...

  12. Frontal Control Process in Intentional Forgetting: Electrophysiological Evidence

    Directory of Open Access Journals (Sweden)

    Heming Gao

    2018-01-01

    Full Text Available In this study, we aimed to seek for the neural evidence of the inhibition control process in directed forgetting (DF. We adopted a modified item-method DF paradigm, in which four kinds of cues were involved. In some trials, the words were followed by only a forgetting (F cue. In the other trials, after a word was presented, a maintenance (M cue was presented, followed by an explicit remembering (M-R cue or an forgetting (M-F cue. Data from 19 healthy adult participants showed that, (1 compared with the remembering cue (i.e., M-R cue, forgetting cues (i.e., M-F cue and F cue evoked enhanced frontal N2 and reduced parietal P3 and late positive complex (LPC components, indicating that the forgetting cues might trigger a more intensive cognitive control process and that fewer amounts of cognitive resources were recruited for the further rehearsal process. (2 Both the M cue and the F cue evoked enhanced N2 and decreased P3 and LPC components than the M-R or M-F cue. These results might indicate that compared with the M-R and M-F cues, both the M and F cues evoked a more intensive cognitive control process and decreased attentional resource allocation process. (3 The F cue evoked a decreased P2 component and an enhanced N2 component relative to the other cues (i.e., M-R, M-F, M, indicating that the F cue received fewer amounts of attentional resources and evoked a more intensive cognitive control process. Taken together, forgetting cues were associated with enhanced N2 activity relative to the maintenance rehearsal process or the remembering process, suggesting an enhanced cognitive control process under DF. This cognitive control process might reflect the role of inhibition in DF as attempting to suppress the ongoing encoding.

  13. A comprehensive analysis of the IMRT dose delivery process using statistical process control (SPC)

    Energy Technology Data Exchange (ETDEWEB)

    Gerard, Karine; Grandhaye, Jean-Pierre; Marchesi, Vincent; Kafrouni, Hanna; Husson, Francois; Aletti, Pierre [Research Center for Automatic Control (CRAN), Nancy University, CNRS, 54516 Vandoeuvre-les-Nancy (France); Department of Medical Physics, Alexis Vautrin Cancer Center, 54511 Vandoeuvre-les-Nancy Cedex (France) and DOSIsoft SA, 94230 Cachan (France); Research Laboratory for Innovative Processes (ERPI), Nancy University, EA 3767, 5400 Nancy Cedex (France); Department of Medical Physics, Alexis Vautrin Cancer Center, 54511 Vandoeuvre-les-Nancy Cedex (France); DOSIsoft SA, 94230 Cachan (France); Research Center for Automatic Control (CRAN), Nancy University, CNRS, 54516 Vandoeuvre-les-Nancy, France and Department of Medical Physics, Alexis Vautrin Cancer Center, 54511 Vandoeuvre-les-Nancy Cedex (France)

    2009-04-15

    The aim of this study is to introduce tools to improve the security of each IMRT patient treatment by determining action levels for the dose delivery process. To achieve this, the patient-specific quality control results performed with an ionization chamber--and which characterize the dose delivery process--have been retrospectively analyzed using a method borrowed from industry: Statistical process control (SPC). The latter consisted in fulfilling four principal well-structured steps. The authors first quantified the short term variability of ionization chamber measurements regarding the clinical tolerances used in the cancer center ({+-}4% of deviation between the calculated and measured doses) by calculating a control process capability (C{sub pc}) index. The C{sub pc} index was found superior to 4, which implies that the observed variability of the dose delivery process is not biased by the short term variability of the measurement. Then, the authors demonstrated using a normality test that the quality control results could be approximated by a normal distribution with two parameters (mean and standard deviation). Finally, the authors used two complementary tools--control charts and performance indices--to thoroughly analyze the IMRT dose delivery process. Control charts aim at monitoring the process over time using statistical control limits to distinguish random (natural) variations from significant changes in the process, whereas performance indices aim at quantifying the ability of the process to produce data that are within the clinical tolerances, at a precise moment. The authors retrospectively showed that the analysis of three selected control charts (individual value, moving-range, and EWMA control charts) allowed efficient drift detection of the dose delivery process for prostate and head-and-neck treatments before the quality controls were outside the clinical tolerances. Therefore, when analyzed in real time, during quality controls, they should

  14. A comprehensive analysis of the IMRT dose delivery process using statistical process control (SPC).

    Science.gov (United States)

    Gérard, Karine; Grandhaye, Jean-Pierre; Marchesi, Vincent; Kafrouni, Hanna; Husson, François; Aletti, Pierre

    2009-04-01

    The aim of this study is to introduce tools to improve the security of each IMRT patient treatment by determining action levels for the dose delivery process. To achieve this, the patient-specific quality control results performed with an ionization chamber--and which characterize the dose delivery process--have been retrospectively analyzed using a method borrowed from industry: Statistical process control (SPC). The latter consisted in fulfilling four principal well-structured steps. The authors first quantified the short-term variability of ionization chamber measurements regarding the clinical tolerances used in the cancer center (+/- 4% of deviation between the calculated and measured doses) by calculating a control process capability (C(pc)) index. The C(pc) index was found superior to 4, which implies that the observed variability of the dose delivery process is not biased by the short-term variability of the measurement. Then, the authors demonstrated using a normality test that the quality control results could be approximated by a normal distribution with two parameters (mean and standard deviation). Finally, the authors used two complementary tools--control charts and performance indices--to thoroughly analyze the IMRT dose delivery process. Control charts aim at monitoring the process over time using statistical control limits to distinguish random (natural) variations from significant changes in the process, whereas performance indices aim at quantifying the ability of the process to produce data that are within the clinical tolerances, at a precise moment. The authors retrospectively showed that the analysis of three selected control charts (individual value, moving-range, and EWMA control charts) allowed efficient drift detection of the dose delivery process for prostate and head-and-neck treatments before the quality controls were outside the clinical tolerances. Therefore, when analyzed in real time, during quality controls, they should improve the

  15. Advanced Map For Real-Time Process Control

    Science.gov (United States)

    Shiobara, Yasuhisa; Matsudaira, Takayuki; Sashida, Yoshio; Chikuma, Makoto

    1987-10-01

    MAP, a communications protocol for factory automation proposed by General Motors [1], has been accepted by users throughout the world and is rapidly becoming a user standard. In fact, it is now a LAN standard for factory automation. MAP is intended to interconnect different devices, such as computers and programmable devices, made by different manufacturers, enabling them to exchange information. It is based on the OSI intercomputer com-munications protocol standard under development by the ISO. With progress and standardization, MAP is being investigated for application to process control fields other than factory automation [2]. The transmission response time of the network system and centralized management of data exchanged with various devices for distributed control are import-ant in the case of a real-time process control with programmable controllers, computers, and instruments connected to a LAN system. MAP/EPA and MINI MAP aim at reduced overhead in protocol processing and enhanced transmission response. If applied to real-time process control, a protocol based on point-to-point and request-response transactions limits throughput and transmission response. This paper describes an advanced MAP LAN system applied to real-time process control by adding a new data transmission control that performs multicasting communication voluntarily and periodically in the priority order of data to be exchanged.

  16. Touch-sensitive graphics terminal applied to process control

    International Nuclear Information System (INIS)

    Bennion, S.I.; Creager, J.D.; VanHouten, R.D.

    1981-01-01

    Limited initial demonstrations of the system described took place during September 1980. A single CRT was used an an input device in the control center while operating a furnace and a pellet inspection gage. These two process line devices were completely controlled, despite the longer than desired response times noted, using a single control station located in the control center. The operator could conveniently execute any function from this remote location which could be performed locally at the hard-wired control panels. With the installation of the enhancements, the integrated touchscreen/graphics terminal will provide a preferable alternative to normal keyboard command input devices

  17. Process controls for radiation hardened aluminum gate bulk silicon CMOS

    International Nuclear Information System (INIS)

    Gregory, B.L.

    1975-01-01

    Optimized dry oxides have recently yielded notable improvements in CMOS radiation-hardness. By following the proper procedures and recipes, it is now possible to produce devices which will function satisfactorily after exposure to a total ionizing dose in excess of 10 6 RADS (Si). This paper is concerned with the controls required on processing parameters once the optimized process is defined. In this process, the pre-irradiation electrical parameters must be closely controlled to insure that devices will function after irradiation. In particular, the specifications on n- and p-channel threshold voltages require tight control of fixed oxide charge, surface-state density, oxide thickness, and substrate and p-well surface concentrations. In order to achieve the above level of radiation hardness, certain processing procedures and parameters must also be closely controlled. Higher levels of cleanliness are required in the hardened process than are commonly required for commercial CMOS since, for hardened dry oxides, no impurity gettering can be employed during or after oxidation. Without such gettering, an unclean oxide is unacceptable due to bias-temperature instability. Correct pre-oxidation cleaning, residual surface damage removal, proper oxidation and annealing temperatures and times, and the correct metal sintering cycle are all important in determining device hardness. In a reproducible, hardened process, each of these processing steps must be closely controlled. (U.S.)

  18. Development of Chemical Process Design and Control for Sustainability

    Directory of Open Access Journals (Sweden)

    Shuyun Li

    2016-07-01

    Full Text Available This contribution describes a novel process systems engineering framework that couples advanced control with sustainability evaluation for the optimization of process operations to minimize environmental impacts associated with products, materials and energy. The implemented control strategy combines a biologically-inspired method with optimal control concepts for finding more sustainable operating trajectories. The sustainability assessment of process operating points is carried out by using the U.S. EPA’s Gauging Reaction Effectiveness for the ENvironmental Sustainability of Chemistries with a multi-Objective Process Evaluator (GREENSCOPE tool that provides scores for the selected indicators in the economic, material efficiency, environmental and energy areas. The indicator scores describe process performance on a sustainability measurement scale, effectively determining which operating point is more sustainable if there are more than several steady states for one specific product manufacturing. Through comparisons between a representative benchmark and the optimal steady states obtained through the implementation of the proposed controller, a systematic decision can be made in terms of whether the implementation of the controller is moving the process towards a more sustainable operation. The effectiveness of the proposed framework is illustrated through a case study of a continuous fermentation process for fuel production, whose material and energy time variation models are characterized by multiple steady states and oscillatory conditions.

  19. Development of Chemical Process Design and Control for ...

    Science.gov (United States)

    This contribution describes a novel process systems engineering framework that couples advanced control with sustainability evaluation and decision making for the optimization of process operations to minimize environmental impacts associated with products, materials, and energy. The implemented control strategy combines a biologically inspired method with optimal control concepts for finding more sustainable operating trajectories. The sustainability assessment of process operating points is carried out by using the U.S. E.P.A.’s Gauging Reaction Effectiveness for the ENvironmental Sustainability of Chemistries with a multi-Objective Process Evaluator (GREENSCOPE) tool that provides scores for the selected indicators in the economic, material efficiency, environmental and energy areas. The indicator scores describe process performance on a sustainability measurement scale, effectively determining which operating point is more sustainable if there are more than several steady states for one specific product manufacturing. Through comparisons between a representative benchmark and the optimal steady-states obtained through implementation of the proposed controller, a systematic decision can be made in terms of whether the implementation of the controller is moving the process towards a more sustainable operation. The effectiveness of the proposed framework is illustrated through a case study of a continuous fermentation process for fuel production, whose materi

  20. Anaerobic Digestion. Student Manual. Biological Treatment Process Control.

    Science.gov (United States)

    Carnegie, John W., Ed.

    This student manual contains the textual material for a four-lesson unit on anaerobic digestion control. Areas addressed include: (1) anaerobic sludge digestion (considering the nature of raw sludge, purposes of anaerobic digestion, the results of digestion, types of equipment, and other topics); (2) digester process control (considering feeding…

  1. The process matters: cyber security in industrial control systems

    NARCIS (Netherlands)

    Hadziosmanovic, D.

    2014-01-01

    An industrial control system (ICS) is a computer system that controls industrial processes such as power plants, water and gas distribution, food production, etc. Since cyber-attacks on an ICS may have devastating consequences on human lives and safety in general, the security of ICS is important.

  2. Statistical Process Control: Going to the Limit for Quality.

    Science.gov (United States)

    Training, 1987

    1987-01-01

    Defines the concept of statistical process control, a quality control method used especially in manufacturing. Generally, concept users set specific standard levels that must be met. Makes the point that although employees work directly with the method, management is responsible for its success within the plant. (CH)

  3. Short Horizon Control Strategies for an Alternating Activated Sludge Process

    DEFF Research Database (Denmark)

    Isaacs, Steven Howard

    1996-01-01

    Three control strategies allowing improved operational flexibility of an alternating type activated sludge process are presented in a unified model based framework. The control handles employed are the addition rate of an external carbon source to denitrification, the cycle length, and the dissol...

  4. Materials control and accountability at the Idaho Chemical Processing Plant

    International Nuclear Information System (INIS)

    Denning, G.E.; Britschgi, J.J.; Spraktes, F.W.

    1985-01-01

    The ICPP high enriched uranium recovery process has historically been operated as a single Material Balance Area (MBA), with input and output measurement capabilities. Safeguards initiated changes in the last five years have resulted in significant materials control and accountability improvements. Those changes include semi-automation of process accountability measurement, data collection and recording; definition of Sub-MBAs; standard plant cleanouts; and, bimonthly inventory estimates. Process monitoring capabilities are also being installed to provide independent operational procedural compliance verification, process anomaly detection, and enhanced materials traceability. Development of a sensitivity analysis approach to defining process measurement requirements is in progress

  5. Automatized material and radioactivity flow control tool in decommissioning process

    International Nuclear Information System (INIS)

    Rehak, I.; Vasko, M.; Daniska, V.; Schultz, O.

    2009-01-01

    In this presentation the automatized material and radioactivity flow control tool in decommissioning process is discussed. It is concluded that: computer simulation of the decommissioning process is one of the important attributes of computer code Omega; one of the basic tools of computer optimisation of decommissioning waste processing are the tools of integral material and radioactivity flow; all the calculated parameters of materials are stored in each point of calculation process and they can be viewed; computer code Omega represents opened modular system, which can be improved; improvement of the module of optimisation of decommissioning waste processing will be performed in the frame of improvement of material procedures and scenarios.

  6. REDUCING WASTEWATER FROM CUCUMBER PICKLING PROCESS BY CONTROLLED CULTURE FERMENTATION

    Science.gov (United States)

    On a demonstration scale, the controlled culture fermentation process (CCF) developed by the U.S. Food Fermentation Laboratory was compared with the conventional natural fermentation process (NF) in regard to product quality and yield and volume and concentration of wastewaters. ...

  7. Software process improvement: controlling developers, managers or users?

    DEFF Research Database (Denmark)

    Nørbjerg, Jacob

    1999-01-01

    The paper discusses how the latest trend in the management of software development: software process improvement (SPI) may affect user-developer relations. At the outset, SPI concerns the "internal workings" of software organisations, but it may also be interpreted as one way to give the developer...... organisation more control over the development process and the relations with the user organization....

  8. Optimal Control of Beer Fermentation Process Using Differential ...

    African Journals Online (AJOL)

    Optimal Control of Beer Fermentation Process Using Differential Transform Method. ... Journal of Applied Sciences and Environmental Management ... The method of differential transform was used to obtain the solution governing the fermentation process; the system of equation was transformed using the differential ...

  9. 21 CFR 114.80 - Processes and controls.

    Science.gov (United States)

    2010-04-01

    ... scheduled process and maintained in all finished foods. Manufacturing shall be in accordance with the... occur often enough to ensure that the container suitably protects the food from leakage or contamination... 21 Food and Drugs 2 2010-04-01 2010-04-01 false Processes and controls. 114.80 Section 114.80 Food...

  10. Simulation of Simple Controlled Processes with Dead-Time.

    Science.gov (United States)

    Watson, Keith R.; And Others

    1985-01-01

    The determination of closed-loop response of processes containing dead-time is typically not covered in undergraduate process control, possibly because the solution by Laplace transforms requires the use of Pade approximation for dead-time, which makes the procedure lengthy and tedious. A computer-aided method is described which simplifies the…

  11. Metrology and process control: dealing with measurement uncertainty

    Science.gov (United States)

    Potzick, James

    2010-03-01

    Metrology is often used in designing and controlling manufacturing processes. A product sample is processed, some relevant property is measured, and the process adjusted to bring the next processed sample closer to its specification. This feedback loop can be remarkably effective for the complex processes used in semiconductor manufacturing, but there is some risk involved because measurements have uncertainty and product specifications have tolerances. There is finite risk that good product will fail testing or that faulty product will pass. Standard methods for quantifying measurement uncertainty have been presented, but the question arises: how much measurement uncertainty is tolerable in a specific case? Or, How does measurement uncertainty relate to manufacturing risk? This paper looks at some of the components inside this process control feedback loop and describes methods to answer these questions.

  12. Hybrid Control and Verification of a Pulsed Welding Process

    DEFF Research Database (Denmark)

    Wisniewski, Rafal; Larsen, Jesper Abildgaard; Izadi-Zamanabadi, Roozbeh

    Currently systems, which are desired to control, are becoming more and more complex and classical control theory objectives, such as stability or sensitivity, are often not sufficient to cover the control objectives of the systems. In this paper it is shown how the dynamics of a pulsed welding...... process can be reformulated into a timed automaton hybrid setting and subsequently properties such as reachability and deadlock absence is verified by the simulation and verification tool UPPAAL....

  13. Software factory techniques applied to process control at CERN

    OpenAIRE

    Dutour, Mathias D

    2007-01-01

    The CERN Large Hadron Collider (LHC) requires constant monitoring and control of quantities of parameters to guarantee operational conditions. For this purpose, a methodology called UNICOS (UNIfied Industrial COntrols Systems) has been implemented to standardize the design of process control applications. To further accelerate the development of these applications, we migrated our existing UNICOS tooling suite toward a software factory in charge of assembling project, domain and technical inf...

  14. A method of automatic data processing in radiometric control

    International Nuclear Information System (INIS)

    Adonin, V.M.; Gulyukina, N.A.; Nemirov, Yu.V.; Mogil'nitskij, M.I.

    1980-01-01

    Described is the algorithm for automatic data processing in gamma radiography of products. Rapidity due to application of recurrent evaluation is a specific feature of the processing. Experimental data of by-line control are presented. The results obtained have shown the applicability of automatic signal processing to the testing under industrial conditions, which would permit to increase the testing efficiency to eliminate the subjectivism in assessment of testing results and to improve working conditions

  15. Treatment of uranium turning with the controllable oxidizing process

    International Nuclear Information System (INIS)

    Shen Bingyi; Zhang Yonggang; Zhen Huikuan

    1989-02-01

    The concept, procedure and safety measures of the controllable oxidizing for uranium turning is described. The feasibility study on technological process has been made. The process provided several advantages such as: simplicity of operation, no pollution environment, safety, high efficiency and low energy consumption. The process can yield nuclear pure uranium dioxide under making no use of a great number of chemical reagent. It may supply raw material for fluoration and provide a simply method of treatment for safe store of uranium turning

  16. 242-A Campaign 99-1 process control plan; FINAL

    International Nuclear Information System (INIS)

    LE, E.Q.

    1999-01-01

    242-A Evaporator 99-1 will process approximately one million gallons of waste from tank 102-AW in June 1999. The process control Plan provides a general description of activities, which will occur during 242-A Evaporator Campaign 99-1 and to document analyses conducted to demonstrate that 102-AW waste is acceptable for processing. Predict is a registered trademark of Risk Decisions England Corporation, United Kingdom

  17. High level model predictive control for plug-and-play process control with stability guaranty

    DEFF Research Database (Denmark)

    Michelsen, Axel Gottlieb; Stoustrup, Jakob

    2010-01-01

    In this paper a method for designing a stabilizing high level model predictive controller for a hierarchical plug- and-play process is presented. This is achieved by abstracting the lower layers of the controller structure as low order models with uncertainty and by using a robust model predictive...... controller for generating the references for these. A simulation example, in which the actuators in a process control system are changed, is reported to show the potential of this approach for plug and play process control....

  18. A process control software package for the SRS

    International Nuclear Information System (INIS)

    Atkins, V.R.; Poole, D.E.; Rawlinson, W.R.

    1980-03-01

    The development of software to give high level access from application programs for monitoring and control of the Daresbury Synchrotron Radiation Source on a network-wide basis is described. The design and implementation of the control system database, a special supervisor call and and 'executive' type task handling of all process input/output services for the 7/32 (which runs under 05/32-MT), and process control 'device driver' software for the 7/16 (run under L5/16-MT) are included. (UK)

  19. Digital signal processing in power system protection and control

    CERN Document Server

    Rebizant, Waldemar; Wiszniewski, Andrzej

    2011-01-01

    Digital Signal Processing in Power System Protection and Control bridges the gap between the theory of protection and control and the practical applications of protection equipment. Understanding how protection functions is crucial not only for equipment developers and manufacturers, but also for their users who need to install, set and operate the protection devices in an appropriate manner. After introductory chapters related to protection technology and functions, Digital Signal Processing in Power System Protection and Control presents the digital algorithms for signal filtering, followed

  20. Intelligent Transportation Control based on Proactive Complex Event Processing

    Directory of Open Access Journals (Sweden)

    Wang Yongheng

    2016-01-01

    Full Text Available Complex Event Processing (CEP has become the key part of Internet of Things (IoT. Proactive CEP can predict future system states and execute some actions to avoid unwanted states which brings new hope to intelligent transportation control. In this paper, we propose a proactive CEP architecture and method for intelligent transportation control. Based on basic CEP technology and predictive analytic technology, a networked distributed Markov decision processes model with predicting states is proposed as sequential decision model. A Q-learning method is proposed for this model. The experimental evaluations show that this method works well when used to control congestion in in intelligent transportation systems.

  1. Electron backscattering for process control in electron beam welding

    International Nuclear Information System (INIS)

    Ardenne, T. von; Panzer, S.

    1983-01-01

    A number of solutions to the automation of electron beam welding is presented. On the basis of electron backscattering a complex system of process control has been developed. It allows an enlarged imaging of the material's surface, improved adjustment of the beam focusing and definite focus positioning. Furthermore, both manual and automated positioning of the electron beam before and during the welding process has become possible. Monitoring of the welding process for meeting standard welding requirements can be achieved with the aid of a control quantity derived from the results of electronic evaluation of the high-frequency electron backscattering

  2. Production process and quality control for the HTTR fuel

    International Nuclear Information System (INIS)

    Yoshimuta, S.; Suzuki, N.; Kaneko, M.; Fukuda, K.

    1991-01-01

    Development of the production and inspection technology for High Temperature Engineering Test Reactor (HTTR) fuel has been carried out by cooperative work between Japan Atomic Energy Research Institute (JAERI) and Nuclear Fuel Industries, Ltd (NFI). The performance and the quality level of the developed fuel are well established to meet the design requirements of the HTTR. For the commercial scale production of the fuel, statistical quality control and quality assurance must be carefully considered in order to assure the safety of the HTTR. It is also important to produce the fuel under well controlled process condition. To meet these requirements in the production of the HTTR fuel, a new production process and quality control system is to be introduced in the new facilities. The main feature of the system is a computer integrated control system. Process control data at each production stage of products and semi-products are all gathered by terminal computers and processed by a host computer. The processed information is effectively used for the production, quality and accountancy control. With the aid of this system, all the products will be easily traceable from starting materials to final stages and the statistical evaluation of the quality of products becomes more reliable. (author). 8 figs

  3. Controllable unit concept as applied to a hypothetical tritium process

    International Nuclear Information System (INIS)

    Seabaugh, P.W.; Sellers, D.E.; Woltermann, H.A.; Boh, D.R.; Miles, J.C.; Fushimi, F.C.

    1976-01-01

    A methodology (controllable unit accountability) is described that identifies controlling errors for corrective action, locates areas and time frames of suspected diversions, defines time and sensitivity limits of diversion flags, defines the time frame in which pass-through quantities of accountable material and by inference SNM remain controllable and provides a basis for identification of incremental cost associated with purely safeguards considerations. The concept provides a rationale from which measurement variability and specific safeguard criteria can be converted into a numerical value that represents the degree of control or improvement attainable with a specific measurement system or combination of systems. Currently the methodology is being applied to a high-throughput, mixed-oxide fuel fabrication process. The process described is merely used to illustrate a procedure that can be applied to other more pertinent processes

  4. Flexible human machine interface for process diagnostics

    International Nuclear Information System (INIS)

    Reifman, J.; Graham, G.E.; Wei, T.Y.C.; Brown, K.R.; Chin, R.Y.

    1996-01-01

    A flexible human machine interface to design and display graphical and textual process diagnostic information is presented. The system operates on different computer hardware platforms, including PCs under MS Windows and UNIX Workstations under X-Windows, in a client-server architecture. The interface system is customized for specific process applications in a graphical user interface development environment by overlaying the image of the process piping and instrumentation diagram with display objects that are highlighted in color during diagnostic display. Customization of the system is presented for Commonwealth Edison's Braidwood PWR Chemical and Volume Control System with transients simulated by a full-scale operator-training simulator and diagnosed by a computer-based system

  5. An application of neural networks to process and materials control

    International Nuclear Information System (INIS)

    Howell, J.A.; Whiteson, R.

    1991-01-01

    Process control consists of two basic elements: a model of the process and knowledge of the desired control algorithm. In some cases the level of the control algorithm is merely supervisory, as in an alarm-reporting or anomaly-detection system. If the model of the process is known, then a set of equations may often be solved explicitly to provide the control algorithm. Otherwise, the model has to be discovered through empirical studies. Neural networks have properties that make them useful in this application. They can learn (make internal models from experience or observations). The problem of anomaly detection in materials control systems fits well into this general control framework. To successfully model a process with a neutral network, a good set of observables must be chosen. These observables must in some sense adequately span the space of representable events, so that a signature metric can be built for normal operation. In this way, a non-normal event, one that does not fit within the signature, can be detected. In this paper, we discuss the issues involved in applying a neural network model to anomaly detection in materials control systems. These issues include data selection and representation, network architecture, prediction of events, the use of simulated data, and software tools. 10 refs., 4 figs., 1 tab

  6. Advanced Control Synthesis for Reverse Osmosis Water Desalination Processes.

    Science.gov (United States)

    Phuc, Bui Duc Hong; You, Sam-Sang; Choi, Hyeung-Six; Jeong, Seok-Kwon

    2017-11-01

      In this study, robust control synthesis has been applied to a reverse osmosis desalination plant whose product water flow and salinity are chosen as two controlled variables. The reverse osmosis process has been selected to study since it typically uses less energy than thermal distillation. The aim of the robust design is to overcome the limitation of classical controllers in dealing with large parametric uncertainties, external disturbances, sensor noises, and unmodeled process dynamics. The analyzed desalination process is modeled as a multi-input multi-output (MIMO) system with varying parameters. The control system is decoupled using a feed forward decoupling method to reduce the interactions between control channels. Both nominal and perturbed reverse osmosis systems have been analyzed using structured singular values for their stabilities and performances. Simulation results show that the system responses meet all the control requirements against various uncertainties. Finally the reduced order controller provides excellent robust performance, with achieving decoupling, disturbance attenuation, and noise rejection. It can help to reduce the membrane cleanings, increase the robustness against uncertainties, and lower the energy consumption for process monitoring.

  7. Studies of neutron methods for process control and criticality surveillance of fissile material processing facilities

    International Nuclear Information System (INIS)

    Zoltowski, T.

    1988-01-01

    The development of radiochemical processes for fissile material processing and spent fuel handling need new control procedures enabling an improvement of plant throughput. This is strictly related to the implementation of continuous criticality control policy and developing reliable methods for monitoring the reactivity of radiochemical plant operations in presence of the process perturbations. Neutron methods seem to be applicable for fissile material control in some technological facilities. The measurement of epithermal neutron source multiplication with heuristic evaluation of measured data enables surveillance of anomalous reactivity enhancement leading to unsafe states. 80 refs., 47 figs., 33 tabs. (author)

  8. Formal Verification of Effectiveness of Control Activities in Business Processes

    Science.gov (United States)

    Arimoto, Yasuhito; Iida, Shusaku; Futatsugi, Kokichi

    It has been an important issue to deal with risks in business processes for achieving companies' goals. This paper introduces a method for applying a formal method to analysis of risks and control activities in business processes in order to evaluate control activities consistently, exhaustively, and to give us potential to have scientific discussion on the result of the evaluation. We focus on document flows in business activities and control activities and risks related to documents because documents play important roles in business. In our method, document flows including control activities are modeled and it is verified by OTS/CafeOBJ Method that risks about falsification of documents are avoided by control activities in the model. The verification is done by interaction between humans and CafeOBJ system with theorem proving, and it raises potential to discuss the result scientifically because the interaction gives us rigorous reasons why the result is derived from the verification.

  9. High-volume manufacturing device overlay process control

    Science.gov (United States)

    Lee, Honggoo; Han, Sangjun; Woo, Jaeson; Lee, DongYoung; Song, ChangRock; Heo, Hoyoung; Brinster, Irina; Choi, DongSub; Robinson, John C.

    2017-03-01

    Overlay control based on DI metrology of optical targets has been the primary basis for run-to-run process control for many years. In previous work we described a scenario where optical overlay metrology is performed on metrology targets on a high frequency basis including every lot (or most lots) at DI. SEM based FI metrology is performed ondevice in-die as-etched on an infrequent basis. Hybrid control schemes of this type have been in use for many process nodes. What is new is the relative size of the NZO as compared to the overlay spec, and the need to find more comprehensive solutions to characterize and control the size and variability of NZO at the 1x nm node: sampling, modeling, temporal frequency and control aspects, as well as trade-offs between SEM throughput and accuracy.

  10. Process Control Systems in the Chemical Industry: Safety vs. Security

    Energy Technology Data Exchange (ETDEWEB)

    Jeffrey Hahn; Thomas Anderson

    2005-04-01

    Traditionally, the primary focus of the chemical industry has been safety and productivity. However, recent threats to our nation’s critical infrastructure have prompted a tightening of security measures across many different industry sectors. Reducing vulnerabilities of control systems against physical and cyber attack is necessary to ensure the safety, security and effective functioning of these systems. The U.S. Department of Homeland Security has developed a strategy to secure these vulnerabilities. Crucial to this strategy is the Control Systems Security and Test Center (CSSTC) established to test and analyze control systems equipment. In addition, the CSSTC promotes a proactive, collaborative approach to increase industry's awareness of standards, products and processes that can enhance the security of control systems. This paper outlines measures that can be taken to enhance the cybersecurity of process control systems in the chemical sector.

  11. Myoelectric signal processing for control of powered limb prostheses.

    Science.gov (United States)

    Parker, P; Englehart, K; Hudgins, B

    2006-12-01

    Progress in myoelectric control technology has over the years been incremental, due in part to the alternating focus of the R&D between control methodology and device hardware. The technology has over the past 50 years or so moved from single muscle control of a single prosthesis function to muscle group activity control of multifunction prostheses. Central to these changes have been developments in the means of extracting information from the myoelectric signal. This paper gives an overview of the myoelectric signal processing challenge, a brief look at the challenge from an historical perspective, the state-of-the-art in myoelectric signal processing for prosthesis control, and an indication of where this field is heading. The paper demonstrates that considerable progress has been made in providing clients with useful and reliable myoelectric communication channels, and that exciting work and developments are on the horizon.

  12. Control of DWPF [Defense Waste Processing Facility] melter feed composition

    International Nuclear Information System (INIS)

    Edwards, R.E. Jr.; Brown, K.G.; Postles, R.L.

    1990-01-01

    The Defense Waste Processing Facility will be used to immobilize Savannah River Site high-level waste into a stable borosilicate glass for disposal in a geologic repository. Proper control of the melter feed composition in this facility is essential to the production of glass which meets product durability constraints dictated by repository regulations and facility processing constraints dictated by melter design. A technique has been developed which utilizes glass property models to determine acceptable processing regions based on the multiple constraints imposed on the glass product and to display these regions graphically. This system along with the batch simulation of the process is being used to form the basis for the statistical process control system for the facility. 13 refs., 3 figs., 1 tab

  13. Flexible distributed architecture for semiconductor process control and experimentation

    Science.gov (United States)

    Gower, Aaron E.; Boning, Duane S.; McIlrath, Michael B.

    1997-01-01

    Semiconductor fabrication requires an increasingly expensive and integrated set of tightly controlled processes, driving the need for a fabrication facility with fully computerized, networked processing equipment. We describe an integrated, open system architecture enabling distributed experimentation and process control for plasma etching. The system was developed at MIT's Microsystems Technology Laboratories and employs in-situ CCD interferometry based analysis in the sensor-feedback control of an Applied Materials Precision 5000 Plasma Etcher (AME5000). Our system supports accelerated, advanced research involving feedback control algorithms, and includes a distributed interface that utilizes the internet to make these fabrication capabilities available to remote users. The system architecture is both distributed and modular: specific implementation of any one task does not restrict the implementation of another. The low level architectural components include a host controller that communicates with the AME5000 equipment via SECS-II, and a host controller for the acquisition and analysis of the CCD sensor images. A cell controller (CC) manages communications between these equipment and sensor controllers. The CC is also responsible for process control decisions; algorithmic controllers may be integrated locally or via remote communications. Finally, a system server images connections from internet/intranet (web) based clients and uses a direct link with the CC to access the system. Each component communicates via a predefined set of TCP/IP socket based messages. This flexible architecture makes integration easier and more robust, and enables separate software components to run on the same or different computers independent of hardware or software platform.

  14. Applying Trusted Network Technology To Process Control Systems

    Science.gov (United States)

    Okhravi, Hamed; Nicol, David

    Interconnections between process control networks and enterprise networks expose instrumentation and control systems and the critical infrastructure components they operate to a variety of cyber attacks. Several architectural standards and security best practices have been proposed for industrial control systems. However, they are based on older architectures and do not leverage the latest hardware and software technologies. This paper describes new technologies that can be applied to the design of next generation security architectures for industrial control systems. The technologies are discussed along with their security benefits and design trade-offs.

  15. Digital signal processing in power electronics control circuits

    CERN Document Server

    Sozanski, Krzysztof

    2013-01-01

    Many digital control circuits in current literature are described using analog transmittance. This may not always be acceptable, especially if the sampling frequency and power transistor switching frequencies are close to the band of interest. Therefore, a digital circuit is considered as a digital controller rather than an analog circuit. This helps to avoid errors and instability in high frequency components. Digital Signal Processing in Power Electronics Control Circuits covers problems concerning the design and realization of digital control algorithms for power electronics circuits using

  16. Fermentation process using specific oxygen uptake rates as a process control

    Science.gov (United States)

    Van Hoek, Pim [Minnetonka, MN; Aristidou, Aristos [Maple Grove, MN; Rush, Brian [Minneapolis, MN

    2011-05-10

    Specific oxygen uptake (OUR) is used as a process control parameter in fermentation processes. OUR is determined during at least the production phase of a fermentation process, and process parameters are adjusted to maintain the OUR within desired ranges. The invention is particularly applicable when the fermentation is conducted using a microorganism having a natural PDC pathway that has been disrupted so that it no longer functions. Microorganisms of this sort often produce poorly under strictly anaerobic conditions. Microaeration controlled by monitoring OUR allows the performance of the microorganism to be optimized.

  17. Use of neural networks in process engineering. Thermodynamics, diffusion, and process control and simulation applications

    International Nuclear Information System (INIS)

    Otero, F

    1998-01-01

    This article presents the current status of the use of Artificial Neural Networks (ANNs) in process engineering applications where common mathematical methods do not completely represent the behavior shown by experimental observations, results, and plant operating data. Three examples of the use of ANNs in typical process engineering applications such as prediction of activity in solvent-polymer binary systems, prediction of a surfactant self-diffusion coefficient of micellar systems, and process control and simulation are shown. These examples are important for polymerization applications, enhanced-oil recovery, and automatic process control

  18. Quality control of the documentation process in electronic economic activities

    Directory of Open Access Journals (Sweden)

    Krutova A.S.

    2017-06-01

    Full Text Available It is proved that the main tool that will provide adequate information resources e economic activities of social and economic relations are documenting quality control processes as the basis of global information space. Directions problems as formation evaluation information resources in the process of documentation, namely development tools assess the efficiency of the system components – qualitative assessment; development of mathematical modeling tools – quantitative evaluation. A qualitative assessment of electronic documentation of economic activity through exercise performance, efficiency of communication; document management efficiency; effectiveness of flow control operations; relationship management effectiveness. The concept of quality control process documents electronically economic activity to components which include: the level of workflow; forms adequacy of information; consumer quality documents; quality attributes; type of income data; condition monitoring systems; organizational level process documentation; attributes of quality, performance quality consumer; type of management system; type of income data; condition monitoring systems. Grounded components of the control system electronic document subjects of economic activity. Detected components IT-audit management system economic activity: compliance audit; audit of internal control; detailed multilevel analysis; corporate risk assessment methodology. The stages and methods of processing electronic transactions economic activity during condition monitoring of electronic economic activity.

  19. Errors in patient specimen collection: application of statistical process control.

    Science.gov (United States)

    Dzik, Walter Sunny; Beckman, Neil; Selleng, Kathleen; Heddle, Nancy; Szczepiorkowski, Zbigniew; Wendel, Silvano; Murphy, Michael

    2008-10-01

    Errors in the collection and labeling of blood samples for pretransfusion testing increase the risk of transfusion-associated patient morbidity and mortality. Statistical process control (SPC) is a recognized method to monitor the performance of a critical process. An easy-to-use SPC method was tested to determine its feasibility as a tool for monitoring quality in transfusion medicine. SPC control charts were adapted to a spreadsheet presentation. Data tabulating the frequency of mislabeled and miscollected blood samples from 10 hospitals in five countries from 2004 to 2006 were used to demonstrate the method. Control charts were produced to monitor process stability. The participating hospitals found the SPC spreadsheet very suitable to monitor the performance of the sample labeling and collection and applied SPC charts to suit their specific needs. One hospital monitored subcategories of sample error in detail. A large hospital monitored the number of wrong-blood-in-tube (WBIT) events. Four smaller-sized facilities, each following the same policy for sample collection, combined their data on WBIT samples into a single control chart. One hospital used the control chart to monitor the effect of an educational intervention. A simple SPC method is described that can monitor the process of sample collection and labeling in any hospital. SPC could be applied to other critical steps in the transfusion processes as a tool for biovigilance and could be used to develop regional or national performance standards for pretransfusion sample collection. A link is provided to download the spreadsheet for free.

  20. Online sensing and control of oil in process wastewater

    Science.gov (United States)

    Khomchenko, Irina B.; Soukhomlinoff, Alexander D.; Mitchell, T. F.; Selenow, Alexander E.

    2002-02-01

    Industrial processes, which eliminate high concentration of oil in their waste stream, find it extremely difficult to measure and control the water purification process. Most oil separation processes involve chemical separation using highly corrosive caustics, acids, surfactants, and emulsifiers. Included in the output of this chemical treatment process are highly adhesive tar-like globules, emulsified and surface oils, and other emulsified chemicals, in addition to suspended solids. The level of oil/hydrocarbons concentration in the wastewater process may fluctuate from 1 ppm to 10,000 ppm, depending upon the specifications of the industry and level of water quality control. The authors have developed a sensing technology, which provides the accuracy of scatter/absorption sensing in a contactless environment by combining these methodologies with reflective measurement. The sensitivity of the sensor may be modified by changing the fluid level control in the flow cell, allowing for a broad range of accurate measurement from 1 ppm to 10,000 ppm. Because this sensing system has been designed to work in a highly invasive environment, it can be placed close to the process source to allow for accurate real time measurement and control.

  1. Process Analytical Technology (PAT): batch-to-batch reproducibility of fermentation processes by robust process operational design and control.

    Science.gov (United States)

    Gnoth, S; Jenzsch, M; Simutis, R; Lübbert, A

    2007-10-31

    The Process Analytical Technology (PAT) initiative of the FDA is a reaction on the increasing discrepancy between current possibilities in process supervision and control of pharmaceutical production processes and its current application in industrial manufacturing processes. With rigid approval practices based on standard operational procedures, adaptations of production reactors towards the state of the art were more or less inhibited for long years. Now PAT paves the way for continuous process and product improvements through improved process supervision based on knowledge-based data analysis, "Quality-by-Design"-concepts, and, finally, through feedback control. Examples of up-to-date implementations of this concept are presented. They are taken from one key group of processes in recombinant pharmaceutical protein manufacturing, the cultivations of genetically modified Escherichia coli bacteria.

  2. Distributed process control system for remote control and monitoring of the TFTR tritium systems

    International Nuclear Information System (INIS)

    Schobert, G.; Arnold, N.; Bashore, D.; Mika, R.; Oliaro, G.

    1989-01-01

    This paper reviews the progress made in the application of a commercially available distributed process control system to support the requirements established for the Tritium REmote Control And Monitoring System (TRECAMS) of the Tokamak Fusion Test REactor (TFTR). The system that will discussed was purchased from Texas (TI) Instruments Automation Controls Division), previously marketed by Rexnord Automation. It consists of three, fully redundant, distributed process controllers interfaced to over 1800 analog and digital I/O points. The operator consoles located throughout the facility are supported by four Digital Equipment Corporation (DEC) PDP-11/73 computers. The PDP-11/73's and the three process controllers communicate over a fully redundant one megabaud fiber optic network. All system functionality is based on a set of completely integrated databases loaded to the process controllers and the PDP-11/73's. (author). 2 refs.; 2 figs

  3. Adaptive model predictive process control using neural networks

    Science.gov (United States)

    Buescher, K.L.; Baum, C.C.; Jones, R.D.

    1997-08-19

    A control system for controlling the output of at least one plant process output parameter is implemented by adaptive model predictive control using a neural network. An improved method and apparatus provides for sampling plant output and control input at a first sampling rate to provide control inputs at the fast rate. The MPC system is, however, provided with a network state vector that is constructed at a second, slower rate so that the input control values used by the MPC system are averaged over a gapped time period. Another improvement is a provision for on-line training that may include difference training, curvature training, and basis center adjustment to maintain the weights and basis centers of the neural in an updated state that can follow changes in the plant operation apart from initial off-line training data. 46 figs.

  4. Analysis of buffering process of control rod hydraulic absorber

    International Nuclear Information System (INIS)

    Bao Jishi; Qin Benke; Bo Hanliang

    2011-01-01

    Control Rod Hydraulic Drive Mechanism(CRHDM) is a newly invented build-in control rod drive mechanism. Hydraulic absorber is the key part of this mechanism, and is used to cushion the control rod when the rod scrams. Thus, it prevents the control rod from being deformed and damaged. In this paper dynamics program ANSYS CFX is used to calculate all kinds of flow conditions in hydraulic absorber to obtain its hydraulic characteristics. Based on the flow resistance coefficients obtained from the simulation results, fluid mass and momentum equations were developed to get the trend of pressure change in the hydraulic cylinder and the displacement of the piston rod during the buffering process of the control rod. The results obtained in this paper indicate that the hydraulic absorber meets the design requirement. The work in this paper will be helpful for the design and optimization of the control rod hydraulic absorber. (author)

  5. Effects of wireless packet loss in industrial process control systems.

    Science.gov (United States)

    Liu, Yongkang; Candell, Richard; Moayeri, Nader

    2017-05-01

    Timely and reliable sensing and actuation control are essential in networked control. This depends on not only the precision/quality of the sensors and actuators used but also on how well the communications links between the field instruments and the controller have been designed. Wireless networking offers simple deployment, reconfigurability, scalability, and reduced operational expenditure, and is easier to upgrade than wired solutions. However, the adoption of wireless networking has been slow in industrial process control due to the stochastic and less than 100% reliable nature of wireless communications and lack of a model to evaluate the effects of such communications imperfections on the overall control performance. In this paper, we study how control performance is affected by wireless link quality, which in turn is adversely affected by severe propagation loss in harsh industrial environments, co-channel interference, and unintended interference from other devices. We select the Tennessee Eastman Challenge Model (TE) for our study. A decentralized process control system, first proposed by N. Ricker, is adopted that employs 41 sensors and 12 actuators to manage the production process in the TE plant. We consider the scenario where wireless links are used to periodically transmit essential sensor measurement data, such as pressure, temperature and chemical composition to the controller as well as control commands to manipulate the actuators according to predetermined setpoints. We consider two models for packet loss in the wireless links, namely, an independent and identically distributed (IID) packet loss model and the two-state Gilbert-Elliot (GE) channel model. While the former is a random loss model, the latter can model bursty losses. With each channel model, the performance of the simulated decentralized controller using wireless links is compared with the one using wired links providing instant and 100% reliable communications. The sensitivity of the

  6. Modified Smith-predictor multirate control utilizing secondary process measurements

    Directory of Open Access Journals (Sweden)

    Rolf Ergon

    2007-01-01

    Full Text Available The Smith-predictor is a well-known control structure for industrial time delay systems, where the basic idea is to estimate the non-delayed process output by use of a process model, and to use this estimate in an inner feedback control loop combined with an outer feedback loop based on the delayed estimation error. The model used may be either mechanistic or identified from input-output data. The paper discusses improvements of the Smith-predictor for systems where also secondary process measurements without time delay are available as a basis for the primary output estimation. The estimator may then be identified also in the common case with primary outputs sampled at a lower rate than the secondary outputs. A simulation example demonstrates the feasibility and advantages of the suggested control structure.

  7. Can we (control) Engineer the degree learning process?

    Science.gov (United States)

    White, A. S.; Censlive, M.; Neilsen, D.

    2014-07-01

    This paper investigates how control theory could be applied to learning processes in engineering education. The initial point for the analysis is White's Double Loop learning model of human automation control modified for the education process where a set of governing principals is chosen, probably by the course designer. After initial training the student decides unknowingly on a mental map or model. After observing how the real world is behaving, a strategy to achieve the governing variables is chosen and a set of actions chosen. This may not be a conscious operation, it maybe completely instinctive. These actions will cause some consequences but not until a certain time delay. The current model is compared with the work of Hollenbeck on goal setting, Nelson's model of self-regulation and that of Abdulwahed, Nagy and Blanchard at Loughborough who investigated control methods applied to the learning process.

  8. Can we (control) Engineer the degree learning process?

    International Nuclear Information System (INIS)

    White, A S; Censlive, M; Neilsen, D

    2014-01-01

    This paper investigates how control theory could be applied to learning processes in engineering education. The initial point for the analysis is White's Double Loop learning model of human automation control modified for the education process where a set of governing principals is chosen, probably by the course designer. After initial training the student decides unknowingly on a mental map or model. After observing how the real world is behaving, a strategy to achieve the governing variables is chosen and a set of actions chosen. This may not be a conscious operation, it maybe completely instinctive. These actions will cause some consequences but not until a certain time delay. The current model is compared with the work of Hollenbeck on goal setting, Nelson's model of self-regulation and that of Abdulwahed, Nagy and Blanchard at Loughborough who investigated control methods applied to the learning process

  9. Closed-Loop Process Control for Electron Beam Freeform Fabrication and Deposition Processes

    Science.gov (United States)

    Taminger, Karen M. (Inventor); Hafley, Robert A. (Inventor); Martin, Richard E. (Inventor); Hofmeister, William H. (Inventor)

    2013-01-01

    A closed-loop control method for an electron beam freeform fabrication (EBF(sup 3)) process includes detecting a feature of interest during the process using a sensor(s), continuously evaluating the feature of interest to determine, in real time, a change occurring therein, and automatically modifying control parameters to control the EBF(sup 3) process. An apparatus provides closed-loop control method of the process, and includes an electron gun for generating an electron beam, a wire feeder for feeding a wire toward a substrate, wherein the wire is melted and progressively deposited in layers onto the substrate, a sensor(s), and a host machine. The sensor(s) measure the feature of interest during the process, and the host machine continuously evaluates the feature of interest to determine, in real time, a change occurring therein. The host machine automatically modifies control parameters to the EBF(sup 3) apparatus to control the EBF(sup 3) process in a closed-loop manner.

  10. Performance of Globally Linearized Controller and Two Region Fuzzy Logic Controller on a Nonlinear Process

    Directory of Open Access Journals (Sweden)

    N. Jaya

    2008-10-01

    Full Text Available In this work, a design and implementation of a Conventional PI controller, single region fuzzy logic controller, two region fuzzy logic controller and Globally Linearized Controller (GLC for a two capacity interacting nonlinear process is carried out. The performance of this process using single region FLC, two region FLC and GLC are compared with the performance of conventional PI controller about an operating point of 50 %. It has been observed that GLC and two region FLC provides better performance. Further, this procedure is also validated by real time experimentation using dSPACE.

  11. Myoelectric control of artificial limb inspired by quantum information processing

    International Nuclear Information System (INIS)

    Siomau, Michael; Jiang, Ning

    2015-01-01

    Precise and elegant coordination of a prosthesis across many degrees of freedom represents a significant challenge to efficient rehabilitation of people with limb deficiency. Processing the electrical neural signals collected from the surface of the remnant muscles of the stump is a common way to initiate and control the different movements available to the artificial limb. Based on the assumption that there are distinguishable and repeatable signal patterns among different types of muscular activation, the problem of prosthesis control reduces to one of pattern recognition. Widely accepted classical methods for pattern recognition, however, cannot provide simultaneous and proportional control of the artificial limb. Here we show that, in principle, quantum information processing of the neural signals allows us to overcome the above-mentioned difficulties, suggesting a very simple scheme for myoelectric control of artificial limb with advanced functionalities. (paper)

  12. Modelling and control of dynamic systems using gaussian process models

    CERN Document Server

    Kocijan, Juš

    2016-01-01

    This monograph opens up new horizons for engineers and researchers in academia and in industry dealing with or interested in new developments in the field of system identification and control. It emphasizes guidelines for working solutions and practical advice for their implementation rather than the theoretical background of Gaussian process (GP) models. The book demonstrates the potential of this recent development in probabilistic machine-learning methods and gives the reader an intuitive understanding of the topic. The current state of the art is treated along with possible future directions for research. Systems control design relies on mathematical models and these may be developed from measurement data. This process of system identification, when based on GP models, can play an integral part of control design in data-based control and its description as such is an essential aspect of the text. The background of GP regression is introduced first with system identification and incorporation of prior know...

  13. A Combined Control Chart for Identifying Out–Of–Control Points in Multivariate Processes

    Directory of Open Access Journals (Sweden)

    Marroquín–Prado E.

    2010-10-01

    Full Text Available The Hotelling's T2 control chart is widely used to identify out–of–control signals in multivariate processes. However, this chart is not sensitive to small shifts in the process mean vec tor. In this work we propose a control chart to identify out–of–control signals. The proposed chart is a combination of Hotelling's T2 chart, M chart proposed by Hayter et al. (1994 and a new chart based on Principal Components. The combination of these charts identifies any type and size of change in the process mean vector. Us ing simulation and the Average Run Length (ARL, the performance of the proposed control chart is evaluated. The ARL means the average points within control before an out–of–control point is detected, The results of the simulation show that the proposed chart is more sensitive that each one of the three charts individually

  14. An introduction to statistical process control in research proteomics.

    Science.gov (United States)

    Bramwell, David

    2013-12-16

    Statistical process control is a well-established and respected method which provides a general purpose, and consistent framework for monitoring and improving the quality of a process. It is routinely used in many industries where the quality of final products is critical and is often required in clinical diagnostic laboratories [1,2]. To date, the methodology has been little utilised in research proteomics. It has been shown to be capable of delivering quantitative QC procedures for qualitative clinical assays [3] making it an ideal methodology to apply to this area of biological research. To introduce statistical process control as an objective strategy for quality control and show how it could be used to benefit proteomics researchers and enhance the quality of the results they generate. We demonstrate that rules which provide basic quality control are easy to derive and implement and could have a major impact on data quality for many studies. Statistical process control is a powerful tool for investigating and improving proteomics research work-flows. The process of characterising measurement systems and defining control rules forces the exploration of key questions that can lead to significant improvements in performance. This work asserts that QC is essential to proteomics discovery experiments. Every experimenter must know the current capabilities of their measurement system and have an objective means for tracking and ensuring that performance. Proteomic analysis work-flows are complicated and multi-variate. QC is critical for clinical chemistry measurements and huge strides have been made in ensuring the quality and validity of results in clinical biochemistry labs. This work introduces some of these QC concepts and works to bridge their use from single analyte QC to applications in multi-analyte systems. This article is part of a Special Issue entitled: Standardization and Quality Control in Proteomics. Copyright © 2013 The Author. Published by Elsevier

  15. A survey of process control computers at the Idaho Chemical Processing Plant

    International Nuclear Information System (INIS)

    Dahl, C.A.

    1989-01-01

    The Idaho Chemical Processing Plant (ICPP) at the Idaho National Engineering Laboratory is charged with the safe processing of spent nuclear fuel elements for the United States Department of Energy. The ICPP was originally constructed in the late 1950s and used state-of-the-art technology for process control at that time. The state of process control instrumentation at the ICPP has steadily improved to keep pace with emerging technology. Today, the ICPP is a college of emerging computer technology in process control with some systems as simple as standalone measurement computers while others are state-of-the-art distributed control systems controlling the operations in an entire facility within the plant. The ICPP has made maximal use of process computer technology aimed at increasing surety, safety, and efficiency of the process operations. Many benefits have been derived from the use of the computers for minimal costs, including decreased misoperations in the facility, and more benefits are expected in the future

  16. Microeconomics of advanced process window control for 50-nm gates

    Science.gov (United States)

    Monahan, Kevin M.; Chen, Xuemei; Falessi, Georges; Garvin, Craig; Hankinson, Matt; Lev, Amir; Levy, Ady; Slessor, Michael D.

    2002-07-01

    Fundamentally, advanced process control enables accelerated design-rule reduction, but simple microeconomic models that directly link the effects of advanced process control to profitability are rare or non-existent. In this work, we derive these links using a simplified model for the rate of profit generated by the semiconductor manufacturing process. We use it to explain why and how microprocessor manufacturers strive to avoid commoditization by producing only the number of dies required to satisfy the time-varying demand in each performance segment. This strategy is realized using the tactic known as speed binning, the deliberate creation of an unnatural distribution of microprocessor performance that varies according to market demand. We show that the ability of APC to achieve these economic objectives may be limited by variability in the larger manufacturing context, including measurement delays and process window variation.

  17. Parallel and distributed processing in power system simulation and control

    Energy Technology Data Exchange (ETDEWEB)

    Falcao, Djalma M [Universidade Federal, Rio de Janeiro, RJ (Brazil). Coordenacao dos Programas de Pos-graduacao de Engenharia

    1994-12-31

    Recent advances in computer technology will certainly have a great impact in the methodologies used in power system expansion and operational planning as well as in real-time control. Parallel and distributed processing are among the new technologies that present great potential for application in these areas. Parallel computers use multiple functional or processing units to speed up computation while distributed processing computer systems are collection of computers joined together by high speed communication networks having many objectives and advantages. The paper presents some ideas for the use of parallel and distributed processing in power system simulation and control. It also comments on some of the current research work in these topics and presents a summary of the work presently being developed at COPPE. (author) 53 refs., 2 figs.

  18. The process of managerial control in quality improvement initiatives.

    Science.gov (United States)

    Slovensky, D J; Fottler, M D

    1994-11-01

    The fundamental intent of strategic management is to position an organization with in its market to exploit organizational competencies and strengths to gain competitive advantage. Competitive advantage may be achieved through such strategies as low cost, high quality, or unique services or products. For health care organizations accredited by the Joint Commission on Accreditation of Healthcare Organizations, continually improving both processes and outcomes of organizational performance--quality improvement--in all operational areas of the organization is a mandated strategy. Defining and measuring quality and controlling the quality improvement strategy remain problematic. The article discusses the nature and processes of managerial control, some potential measures of quality, and related information needs.

  19. Process control with optical emission spectroscopy in triode ion plating

    International Nuclear Information System (INIS)

    Salmenoja, K.; Korhonen, A.S.; Sulonen, M.S.

    1985-01-01

    Physical vapor deposition (PVD) techniques used to prepare, e.g., hard TiN, HfN, or ZrN coatings include a great variety of processes ranging from reactive evaporation to sputtering and ion plating. In ion plating one effective way to enhance ionization is to use a negatively biased hot filament. The use of an electron emitting filament brings an extra variable to be taken into account in developing the process control. In addition, proper control of the evaporation source is critical in ensuring reproducible results. With optical emission spectroscopy (OES) it should be possible to control the coating process more accurately. The stoichiometry and the composition of the growing coating may then be ensured effectively in subsequent runs. In this work the application of optical emission spectroscopy for process control in triode ion plating is discussed. The composition of the growing coating is determined experimentally using the relative intensities of specific emission lines. Changes in the evaporation rate and the gas flow can be seen directly from emission line intensities. Even the so-called poisoning of the evaporation source with reactive gas can be detected. Several experimental runs were carried out and afterwards the concentration profiles of the deposited coatings were checked with the nuclear resonance broadening (NRB) method. The results show the usefulness of emission spectroscopy in discharge control

  20. Intelligent process control of fiber chemical vapor deposition

    Science.gov (United States)

    Jones, John Gregory

    Chemical Vapor Deposition (CVD) is a widely used process for the application of thin films. In this case, CVD is being used to apply a thin film interface coating to single crystal monofilament sapphire (Alsb2Osb3) fibers for use in Ceramic Matrix Composites (CMC's). The hot-wall reactor operates at near atmospheric pressure which is maintained using a venturi pump system. Inert gas seals obviate the need for a sealed system. A liquid precursor delivery system has been implemented to provide precise stoichiometry control. Neural networks have been implemented to create real-time process description models trained using data generated based on a Navier-Stokes finite difference model of the process. Automation of the process to include full computer control and data logging capability is also presented. In situ sensors including a quadrupole mass spectrometer, thermocouples, laser scanner, and Raman spectrometer have been implemented to determine the gas phase reactants and coating quality. A fuzzy logic controller has been developed to regulate either the gas phase or the in situ temperature of the reactor using oxygen flow rate as an actuator. Scanning electron microscope (SEM) images of various samples are shown. A hierarchical control structure upon which the control structure is based is also presented.

  1. Process Control Security in the Cybercrime Information Exchange NICC

    OpenAIRE

    Luiijf, H.A.M.

    2009-01-01

    Detecting, investigating and prosecuting cybercrime? Extremely important, but not really the solution for the problem. Prevention is better! The sectors that have joined the Cybercrime Information Exchange have accepted the challenge of ensuring the effectiveness of the (information) security of process control systems (PCS), including SCADA. This publication makes it clear why it is vital that organizations establish and maintain control over the security of the information and communication...

  2. Chemical sensors and gas sensors for process control in biotechnology

    International Nuclear Information System (INIS)

    Williams, D.E.

    1988-04-01

    This paper is concerned with the possibilities for chemical measurement of the progress of biotechnological processes which are offered by devices already developed for other demanding applications. It considers the potential use of ultrasonic instrumentation originally developed for the nuclear industry, gas measurement methods from the fields of environmental monitoring and combustion control, nuclear instruments developed for the oil, mining and chemical industries, robotic systems and advanced control techniques. (author)

  3. Computer-integrated electric-arc melting process control system

    OpenAIRE

    Дёмин, Дмитрий Александрович

    2014-01-01

    Developing common principles of completing melting process automation systems with hardware and creating on their basis rational choices of computer- integrated electricarc melting control systems is an actual task since it allows a comprehensive approach to the issue of modernizing melting sites of workshops. This approach allows to form the computer-integrated electric-arc furnace control system as part of a queuing system “electric-arc furnace - foundry conveyor” and consider, when taking ...

  4. Quality control of CANDU6 fuel element in fabrication process

    International Nuclear Information System (INIS)

    Li Yinxie; Zhang Jie

    2012-01-01

    To enhance the fine control over all aspects of the production process, improve product quality, fuel element fabrication process for CANDU6 quality process control activities carried out by professional technical and management technology combined mode, the quality of the fuel elements formed around CANDU6 weak links - - end plug , and brazing processes and procedures associated with this aspect of strict control, in improving staff quality consciousness, strengthening equipment maintenance, improved tooling, fixtures, optimization process test, strengthen supervision, fine inspection operations, timely delivery carry out aspects of the quality of information and concerns the production environment, etc., to find the problem from the improvement of product quality and factors affecting the source, and resolved to form the active control, comprehensive and systematic analysis of the problem of the quality management concepts, effectively reducing the end plug weld microstructure after the failure times and number of defects zirconium alloys brazed, improved product quality, and created economic benefits expressly provided, while staff quality consciousness and attention to detail, collaboration department, communication has been greatly improved and achieved very good management effectiveness. (authors)

  5. Exporting Variables in a Hierarchically Distributed Control System

    Energy Technology Data Exchange (ETDEWEB)

    Chamizo Llatas, M

    1995-07-01

    We describe the Remote Variable Access Service (RVAS), a network service developed and used in the distributed control and monitoring system of the TJ-II Heliac, which is under construction at CIEMAT (Madrid, Spain) and devoted to plasma studies in the nuclear fusion field. The architecture of the TJ-II control system consists of one central Sun workstation Sparc 10 and several autonomous subsystems based on VME crates with embedded processors running the OS-9 (V.24) real time operating system. The RVAS service allows state variables in local control processes running in subsystems to be exported to remote processes running in the central control workstation. Thus we extend the concept of exporting of file systems in UNIX machines to variables in processes running in different machines. (Author) 6 refs.

  6. Exporting Variables in a Hierarchically Distributed Control System

    International Nuclear Information System (INIS)

    Diaz Martin; Martinez Laso, L.

    1995-01-01

    We describe the Remote Variable Access Service (RVAS), a network service developed and use in the distributed control and monitoring system of the TJ-II Heliac, which is under construction at CIEMAT (Madrid, Spain) and devoted to plasma studies in the nuclear fusion field. The architecture of the TJ-II control system consists of one central Sun workstation Sparc 10 and several autonomous subsystems based on VME crates with embedded processors running the os-9 (V.24) real time operating system. The RVAS service allows state variables in local control processes running in subsystems to be exported to remote processes running in the central control workstation. Thus we extend the concept of exporting of file systems in UNIX machines to variables in processes running in different machines. (Author)

  7. Exporting Variables in a Hierarchically Distributed Control System

    International Nuclear Information System (INIS)

    Chamizo Llatas, M.

    1995-01-01

    We describe the Remote Variable Access Service (RVAS), a network service developed and used in the distributed control and monitoring system of the TJ-II Heliac, which is under construction at CIEMAT (Madrid, Spain) and devoted to plasma studies in the nuclear fusion field. The architecture of the TJ-II control system consists of one central Sun workstation Sparc 10 and several autonomous subsystems based on VME crates with embedded processors running the OS-9 (V.24) real time operating system. The RVAS service allows state variables in local control processes running in subsystems to be exported to remote processes running in the central control workstation. Thus we extend the concept of exporting of file systems in UNIX machines to variables in processes running in different machines. (Author) 6 refs

  8. Quality Control in Automated Manufacturing Processes – Combined Features for Image Processing

    Directory of Open Access Journals (Sweden)

    B. Kuhlenkötter

    2006-01-01

    Full Text Available In production processes the use of image processing systems is widespread. Hardware solutions and cameras respectively are available for nearly every application. One important challenge of image processing systems is the development and selection of appropriate algorithms and software solutions in order to realise ambitious quality control for production processes. This article characterises the development of innovative software by combining features for an automatic defect classification on product surfaces. The artificial intelligent method Support Vector Machine (SVM is used to execute the classification task according to the combined features. This software is one crucial element for the automation of a manually operated production process

  9. Statistic techniques of process control for MTR type

    International Nuclear Information System (INIS)

    Oliveira, F.S.; Ferrufino, F.B.J.; Santos, G.R.T.; Lima, R.M.

    2002-01-01

    This work aims at introducing some improvements on the fabrication of MTR type fuel plates, applying statistic techniques of process control. The work was divided into four single steps and their data were analyzed for: fabrication of U 3 O 8 fuel plates; fabrication of U 3 Si 2 fuel plates; rolling of small lots of fuel plates; applying statistic tools and standard specifications to perform a comparative study of these processes. (author)

  10. Data processing system for real-time control

    International Nuclear Information System (INIS)

    Oasa, K.; Mochizuki, O.; Toyokawa, R.; Yahiro, K.

    1983-01-01

    Real-time control, for large Tokamak JT-60, requires various data processings between diagnostic devices to control system. These processings require to high speed performance so that it aims at giving information necessary for feedback control during discharges. Then, the architecture of this system has hierachical structure of processors. These processors are connected each other by the CAMAC modules and the optical communication network, which is the 5 M bytes/second CAMAC serial highway. This system has two kinds of intelligences for this purpose. One is ACM-PU pairs in some torus hall crates which has a microcomputerized auxiliary controller and a preprocessing unit. Other is real-time processor which has a minicomputer and preprocessing unit. Most of the real-time processing, for example Abel inversion are characteristic to the diagnostic devices. Such a processing is carried out by an ACM-PU pair in the crate dedicated to the diagnostic device. Some processings, however, are also necessary which compute secondary parameters as functions of primary parameters. A typical example is Zeff, which is a function of Te, Ne and bremsstrahluny intensity. The real-time processor is equipped for such secondary processings and transfer the results. Preprocessing unit -PU- attached to ACM and real-time processor contains a signal processor, which executes in parallel such function as move, add and multiply during one micro-instruction cycle of 200 nsec. According to the progress of the experiment, more high speed processing are required, so the authors developed the PU-X module that contains multi signal processors. After a shot, inter-shot-processor which consists of general-purpose computers, gathers data into the database, then analyze them, and improve these processes to more effective

  11. In-line metallurgical process control in the steel industry

    International Nuclear Information System (INIS)

    Wanin, M.

    1993-01-01

    The steel products manufacturing involves a long line of complex processes: liquid metal elaboration, solidification, hot and cold transformation by rolling surface protection by coating. The Process Control aims at improving global productivity and quality of the resulting products by optimizing each elementary process as well as management of tools or workshops interfaces. Complex processes, involving generally many variables, require for their control more or less sophisticated models. These process models are either analytical when physical and thermodynamical mechanisms are known or statistical or knowledge based, according to circumstances. In any case, it is necessary to have a reliable and precise instrumentation to adjust undetermined parameters during model development and to be able to take into account external parameters variability during current working. This instrumentation concerns both running of machines and testing of manufactured materials under harsh environment conditions of Iron and Steel industry: temperature, dusts, steam, electromagnetic interferences, vibrations, .. . In this context, in-line Non Destructive Testing methods contribute efficienly because they may give directly and in real time products characteristics, integrating both drifts of machines and sensors due to their ageing and the abnormal spread of material entering the process. These methods induce the development of sophisticated inspection equipments whose strategic significance is such that their failure to operate can require production shutdown. The paper gives some representative examples of improvement of the accuracy of an in-line measurement or controlling of elementary processes or processes interfaces: temperature measurement by infrared pyrometry, thickness profile determination by X-ray array sensor, recrystallization control in continuous by X-ray and ultrasonic methods, automatic detection and indentification of surface defects by optics, cracks detection on

  12. A systematic framework for design of process monitoring and control (PAT) systems for crystallization processes

    DEFF Research Database (Denmark)

    Abdul Samad, Noor Asma Fazli Bin; Sin, Gürkan; Gernaey, Krist

    2013-01-01

    A generic computer-aided framework for systematic design of a process monitoring and control system for crystallization processes has been developed to study various aspects of crystallization operations.The systematic design framework contains a generic crystallizer modelling toolbox, a tool for...

  13. Monitoring a PVC batch process with multivariate statistical process control charts

    NARCIS (Netherlands)

    Tates, A. A.; Louwerse, D. J.; Smilde, A. K.; Koot, G. L. M.; Berndt, H.

    1999-01-01

    Multivariate statistical process control charts (MSPC charts) are developed for the industrial batch production process of poly(vinyl chloride) (PVC). With these MSPC charts different types of abnormal batch behavior were detected on-line. With batch contribution plots, the probable causes of these

  14. Artificial Intelligence for Inferential Control of Crude Oil Stripping Process

    Directory of Open Access Journals (Sweden)

    Mehdi Ebnali

    2018-01-01

    Full Text Available Stripper columns are used for sweetening crude oil, and they must hold product hydrogen sulfide content as near the set points as possible in the faces of upsets. Since product    quality cannot be measured easily and economically online, the control of product quality is often achieved by maintaining a suitable tray temperature near its set point. Tray temperature control method, however, is not a proper option for a multi-component stripping column because the tray temperature does not correspond exactly to the product composition. To overcome this problem, secondary measurements can be used to infer the product quality and adjust the values of the manipulated variables. In this paper, we have used a novel inferential control approach base on adaptive network fuzzy inference system (ANFIS for stripping process. ANFIS with different learning algorithms is used for modeling the process and building a composition estimator to estimate the composition of the bottom product. The developed estimator is tested, and the results show that the predictions made by ANFIS structure are in good agreement with the results of simulation by ASPEN HYSYS process simulation package. In addition, inferential control by the implementation of ANFIS-based online composition estimator in a cascade control scheme is superior to traditional tray temperature control method based on less integral time absolute error and low duty consumption in reboiler.

  15. Model Predictive Control of Mineral Column Flotation Process

    Directory of Open Access Journals (Sweden)

    Yahui Tian

    2018-06-01

    Full Text Available Column flotation is an efficient method commonly used in the mineral industry to separate useful minerals from ores of low grade and complex mineral composition. Its main purpose is to achieve maximum recovery while ensuring desired product grade. This work addresses a model predictive control design for a mineral column flotation process modeled by a set of nonlinear coupled heterodirectional hyperbolic partial differential equations (PDEs and ordinary differential equations (ODEs, which accounts for the interconnection of well-stirred regions represented by continuous stirred tank reactors (CSTRs and transport systems given by heterodirectional hyperbolic PDEs, with these two regions combined through the PDEs’ boundaries. The model predictive control considers both optimality of the process operations and naturally present input and state/output constraints. For the discrete controller design, spatially varying steady-state profiles are obtained by linearizing the coupled ODE–PDE model, and then the discrete system is obtained by using the Cayley–Tustin time discretization transformation without any spatial discretization and/or without model reduction. The model predictive controller is designed by solving an optimization problem with input and state/output constraints as well as input disturbance to minimize the objective function, which leads to an online-solvable finite constrained quadratic regulator problem. Finally, the controller performance to keep the output at the steady state within the constraint range is demonstrated by simulation studies, and it is concluded that the optimal control scheme presented in this work makes this flotation process more efficient.

  16. CEBAF control system

    International Nuclear Information System (INIS)

    Bork, R.; Grubb, C.; Lahti, G.; Navarro, E.; Sage, J.

    1989-01-01

    A logic-based computer control system is in development at CEBAF. This Unix/C language software package, running on a distributed, hierarchical system of workstation and supervisory minicomputers, interfaces to hardware via CAMAC. Software aspects to be covered are ladder logic, interactive database generation, networking, and graphic user interfaces. 1 fig

  17. A Survey of Research in Supervisory Control and Data Acquisition (SCADA)

    Science.gov (United States)

    2014-09-01

    RISC ) platforms running some version of UNIX.4 Around the turn of the millennium work began on applying Web technologies to SCADA systems.5–8 Lately...the 2 trend has been to move from the UNIX/ RISC system to commodity hardware and Microsoft solutions although there is some Linux,1 to move from...Control Center MTU Master Terminal Unit OS operating system PKI Public Key Infrastructure PLC Programmable Logic Controller RISC Reduced Instruction

  18. Improving the effectiveness of detailed processing by dynamic control of processing with high sports range

    Directory of Open Access Journals (Sweden)

    Yu.V. Shapoval

    2017-12-01

    Full Text Available In this article the possibility of increasing the efficiency of the processing of parts with a diameter of up to 20 mm is analyzed, namely: vibration resistance of the cutting process at pinching due to cutting speed control in the processing, forecasting and selection of rotational frequencies, which ensure the stability of the processing system, controlling the dynamics of the process of displacement of the additional mass. The method of investigation of vibration processes during the sharpening is developed. As a result of the processing of experimental data, it was found that when an oscillatory motion is applied to the spindle rotation, the overall level of oscillation decreases, which is reflected on the quality of the treated surface. The choice of a previously known spindle rotation frequency range at which the lowest value of the oscillation amplitude of the instrument is observed in the radial direction to the detail part, allows you to increase the processing efficiency while maintaining the drawing requirements for roughness by increasing the spindle rotational speed. The combination of the node of the own forms of oscillation and the cutting zone, by dynamically controlling the fluctuations of the lathe armature due to the increase of the inertia characteristics of the machine and the reduction of the oscillation amplitude of the tool, can improve the accuracy of machining and roughness of the processed surface of the component at higher spindle speeds.

  19. Automatic and controlled processing and the Broad Autism Phenotype.

    Science.gov (United States)

    Camodeca, Amy; Voelker, Sylvia

    2016-01-30

    Research related to verbal fluency in the Broad Autism Phenotype (BAP) is limited and dated, but generally suggests intact abilities in the context of weaknesses in other areas of executive function (Hughes et al., 1999; Wong et al., 2006; Delorme et al., 2007). Controlled processing, the generation of search strategies after initial, automated responses are exhausted (Spat, 2013), has yet to be investigated in the BAP, and may be evidenced in verbal fluency tasks. One hundred twenty-nine participants completed the Delis-Kaplan Executive Function System Verbal Fluency test (D-KEFS; Delis et al., 2001) and the Broad Autism Phenotype Questionnaire (BAPQ; Hurley et al., 2007). The BAP group (n=53) produced significantly fewer total words during the 2nd 15" interval compared to the Non-BAP (n=76) group. Partial correlations indicated similar relations between verbal fluency variables for each group. Regression analyses predicting 2nd 15" interval scores suggested differentiation between controlled and automatic processing skills in both groups. Results suggest adequate automatic processing, but slowed development of controlled processing strategies in the BAP, and provide evidence for similar underlying cognitive constructs for both groups. Controlled processing was predictive of Block Design score for Non-BAP participants, and was predictive of Pragmatic Language score on the BAPQ for BAP participants. These results are similar to past research related to strengths and weaknesses in the BAP, respectively, and suggest that controlled processing strategy use may be required in instances of weak lower-level skills. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  20. Elements for successful sensor-based process control {Integrated Metrology}

    International Nuclear Information System (INIS)

    Butler, Stephanie Watts

    1998-01-01

    Current productivity needs have stimulated development of alternative metrology, control, and equipment maintenance methods. Specifically, sensor applications provide the opportunity to increase productivity, tighten control, reduce scrap, and improve maintenance schedules and procedures. Past experience indicates a complete integrated solution must be provided for sensor-based control to be used successfully in production. In this paper, Integrated Metrology is proposed as the term for an integrated solution that will result in a successful application of sensors for process control. This paper defines and explores the perceived four elements of successful sensor applications: business needs, integration, components, and form. Based upon analysis of existing successful commercially available controllers, the necessary business factors have been determined to be strong, measurable industry-wide business needs whose solution is profitable and feasible. This paper examines why the key aspect of integration is the decision making process. A detailed discussion is provided of the components of most importance to sensor based control: decision-making methods, the 3R's of sensors, and connectivity. A metric for one of the R's (resolution) is proposed to allow focus on this important aspect of measurement. A form for these integrated components which synergistically partitions various aspects of control at the equipment and MES levels to efficiently achieve desired benefits is recommended

  1. Elements for successful sensor-based process control {Integrated Metrology}

    Science.gov (United States)

    Butler, Stephanie Watts

    1998-11-01

    Current productivity needs have stimulated development of alternative metrology, control, and equipment maintenance methods. Specifically, sensor applications provide the opportunity to increase productivity, tighten control, reduce scrap, and improve maintenance schedules and procedures. Past experience indicates a complete integrated solution must be provided for sensor-based control to be used successfully in production. In this paper, Integrated Metrology is proposed as the term for an integrated solution that will result in a successful application of sensors for process control. This paper defines and explores the perceived four elements of successful sensor applications: business needs, integration, components, and form. Based upon analysis of existing successful commercially available controllers, the necessary business factors have been determined to be strong, measurable industry-wide business needs whose solution is profitable and feasible. This paper examines why the key aspect of integration is the decision making process. A detailed discussion is provided of the components of most importance to sensor based control: decision-making methods, the 3R's of sensors, and connectivity. A metric for one of the R's (resolution) is proposed to allow focus on this important aspect of measurement. A form for these integrated components which synergistically partitions various aspects of control at the equipment and MES levels to efficiently achieve desired benefits is recommended.

  2. Knowledge-based processing for aircraft flight control

    Science.gov (United States)

    Painter, John H.; Glass, Emily; Economides, Gregory; Russell, Paul

    1994-01-01

    This Contractor Report documents research in Intelligent Control using knowledge-based processing in a manner dual to methods found in the classic stochastic decision, estimation, and control discipline. Such knowledge-based control has also been called Declarative, and Hybid. Software architectures were sought, employing the parallelism inherent in modern object-oriented modeling and programming. The viewpoint adopted was that Intelligent Control employs a class of domain-specific software architectures having features common over a broad variety of implementations, such as management of aircraft flight, power distribution, etc. As much attention was paid to software engineering issues as to artificial intelligence and control issues. This research considered that particular processing methods from the stochastic and knowledge-based worlds are duals, that is, similar in a broad context. They provide architectural design concepts which serve as bridges between the disparate disciplines of decision, estimation, control, and artificial intelligence. This research was applied to the control of a subsonic transport aircraft in the airport terminal area.

  3. A Taguchi approach on optimal process control parameters for HDPE pipe extrusion process

    Science.gov (United States)

    Sharma, G. V. S. S.; Rao, R. Umamaheswara; Rao, P. Srinivasa

    2017-06-01

    High-density polyethylene (HDPE) pipes find versatile applicability for transportation of water, sewage and slurry from one place to another. Hence, these pipes undergo tremendous pressure by the fluid carried. The present work entails the optimization of the withstanding pressure of the HDPE pipes using Taguchi technique. The traditional heuristic methodology stresses on a trial and error approach and relies heavily upon the accumulated experience of the process engineers for determining the optimal process control parameters. This results in setting up of less-than-optimal values. Hence, there arouse a necessity to determine optimal process control parameters for the pipe extrusion process, which can ensure robust pipe quality and process reliability. In the proposed optimization strategy, the design of experiments (DoE) are conducted wherein different control parameter combinations are analyzed by considering multiple setting levels of each control parameter. The concept of signal-to-noise ratio ( S/ N ratio) is applied and ultimately optimum values of process control parameters are obtained as: pushing zone temperature of 166 °C, Dimmer speed at 08 rpm, and Die head temperature to be 192 °C. Confirmation experimental run is also conducted to verify the analysis and research result and values proved to be in synchronization with the main experimental findings and the withstanding pressure showed a significant improvement from 0.60 to 1.004 Mpa.

  4. Fully Automated Concentration Control of the Acidic Texturisation Process

    OpenAIRE

    Dannenberg, T.; Zimmer, M.; Rentsch, J.

    2012-01-01

    To enable a concentration control in the acidic texturing process we have closed the feedback loop from analytical data to the dosing mechanism of the used process tool. In order to analyze the process bath we used near-infrared spectroscopy in an online setup as well as ion chromatography as an inline method in a second approach. Using the developed dosing algorithm allows a concentration optimization of HF and HNO3 in dependence of the Si concentrations. This allows a further optimization o...

  5. Software factory techniques applied to Process Control at CERN

    CERN Multimedia

    Dutour, MD

    2007-01-01

    The CERN Large Hadron Collider (LHC) requires constant monitoring and control of quantities of parameters to guarantee operational conditions. For this purpose, a methodology called UNICOS (UNIfied Industrial COntrols Systems) has been implemented to standardize the design of process control applications. To further accelerate the development of these applications, we migrated our existing UNICOS tooling suite toward a software factory in charge of assembling project, domain and technical information seamlessly into deployable PLC (Programmable logic Controller) – SCADA (Supervisory Control And Data Acquisition) systems. This software factory delivers consistently high quality by reducing human error and repetitive tasks, and adapts to user specifications in a cost-efficient way. Hence, this production tool is designed to encapsulate and hide the PLC and SCADA target platforms, enabling the experts to focus on the business model rather than specific syntaxes and grammars. Based on industry standard software...

  6. Software factory techniques applied to process control at CERN

    CERN Document Server

    Dutour, Mathias D

    2008-01-01

    The CERN Large Hadron Collider (LHC) requires constant monitoring and control of quantities of parameters to guarantee operational conditions. For this purpose, a methodology called UNICOS (UNIfied Industrial COntrols Systems) has been implemented to standardize the design of process control applications. To further accelerate the development of these applications, we migrated our existing UNICOS tooling suite toward a software factory in charge of assembling project, domain and technical information seamlessly into deployable PLC (Programmable logic Controller) - SCADA (Supervisory Control And Data Acquisition) systems. This software factory delivers consistently high quality by reducing human error and repetitive tasks, and adapts to user specifications in a cost-efficient way. Hence, this production tool is designed to encapsulate and hide the PLC and SCADA target platforms, enabling the experts to focus on the business model rather than specific syntaxes and grammars. Based on industry standard software, ...

  7. Inteligent control system for a CANDU 600 type reactor process

    International Nuclear Information System (INIS)

    Venescu, B.; Zevedei, D.; Jurian, M.; Venescu, R.

    2013-01-01

    The present paper is set on presenting a highly intelligent configuration, capable of controlling, without the need of the human factor, a complete nuclear power plant type of system, giving it the status of an autonomous system. The urge for such a controlling system is justified by the amount of drawbacks that appear in real life as disadvantages, loses and sometimes even inefficiency in the current controlling and comanding systems of the nuclear reactors. The application stands in the comand sent from the auxiliary feedwater flow control valves to the steam generators. As an environment fit for development I chose Matlab Simulink to simulate the behaviour of the process and the adjusted system. Comparing the results obtained after the fuzzy regulation with those obtained after the classical regulation, we can demonstrate the necessity of implementing artificial intelligence techniques in nuclear power plants and we can agree to the advantages of being able to control everything automatically. (authors)

  8. Process automation using combinations of process and machine control technologies with application to a continuous dissolver

    International Nuclear Information System (INIS)

    Spencer, B.B.; Yarbro, O.O.

    1991-01-01

    Operation of a continuous rotary dissolver, designed to leach uranium-plutonium fuel from chopped sections of reactor fuel cladding using nitric acid, has been automated. The dissolver is a partly continuous, partly batch process that interfaces at both ends with batchwise processes, thereby requiring synchronization of certain operations. Liquid acid is fed and flows through the dissolver continuously, whereas chopped fuel elements are fed to the dissolver in small batches and move through the compartments of the dissolver stagewise. Sequential logic (or machine control) techniques are used to control discrete activities such as the sequencing of isolation valves. Feedback control is used to control acid flowrates and temperatures. Expert systems technology is used for on-line material balances and diagnostics of process operation. 1 ref., 3 figs

  9. Project B610 process control configuration acceptance test procedure

    International Nuclear Information System (INIS)

    Silvan, G.R.

    1994-01-01

    The purpose of this test is to verify the Westinghouse configuration of the MICON A/S Distributed Control System for project B610. The following will be verified: proper assignment and operation of all field inputs to and outputs from the MICON Termination panels; proper operation of all display data on the operator's console; proper operation of all required alarms; and proper operation of all required interlocks. The MICON A/S control system is configured to replace all the control, indication, and alarm panels now located in the Power Control Room. Nine systems are covered by this control configuration, 2736-ZB HVAC, 234-5Z HVAC, Process Vacuum, Dry Air, 291-Z Closed Loop Cooling, Building Accelerometer, Evacuation Siren, Stack CAMs, and Fire. The 2736-ZB HVAC system consists of the ventilation controls for 2736-ZB and 2736-Z as well as alarms for the emergency generators and 232-Z. The 234-5Z HVAC system is the ventilation controls for 235-5Z and 236-Z buildings. Process Vacuum covers the controls for the 26 inch vacuum system. Dry Air covers the controls for the steam and electric air dryers. The 291-Z Closed Loop Cooling system consists of the status indications and alarms for the 291-Z compressor and vacuum pump closed loop cooling system. The rest of closed loop cooling was tested earlier. The Building Accelerometer system consists of the status indications for the two seismic system accelerometers. The Evacuation Siren system includes the controls for the evacuation and take cover sirens. Stack CAMs cover the alarms for the various building ventilation stack continuous air monitors. Finally, the Fire system covers the various fire alarms now located in Room 321-A

  10. Development of quality control system for fingerprint comparison processes

    Directory of Open Access Journals (Sweden)

    Shiquan Liu

    2017-01-01

    Full Text Available Fingerprint evidence played an important role in investigation, prosecution, and trial process due to the belief of its uniqueness and unchanged characteristics. However, in recent years, the science behind the process of fingerprint comparisons has been questioned. Main research questions have been focusing on the opaqueness within the comparison processes, subjective judgments, lack of universal standards, no error rate expression on final conclusions, and poor scientific fundamental research data. Facing the above-mentioned questions, this paper aims to suggest a quality control system (QCS for fingerprint comparison processes. This QCS is based on the use of software (PiAnoS and its technological features, being able to provide a data management model to increase the transparency and quality of fingerprint comparison processes.

  11. Quality Control through Design and Process: Gambrel Roof Truss Challenge

    Science.gov (United States)

    Ward, Dell; Jones, James

    2011-01-01

    Customers determine whether a product fulfills their needs or satisfies them. "Quality control", then, is the process of finding out what the customer wants, along with designing, producing, delivering, and servicing the product--and ultimately satisfying the customer's expectations. For many years, people considered a product to be of good…

  12. Exploring control parameters of two photon processes in solutions

    Indian Academy of Sciences (India)

    Here, we present the effect of several control parameters on the TPA process that are independent of .... as the typical selection rules and pathways of mole- cular transitions for ..... Inset in the graph shows the 780 beam spec- tra at two ...

  13. Migration and demographical processes in the zone of radiation control

    International Nuclear Information System (INIS)

    Likhachev, N.E.

    2006-01-01

    The state of demographical processes in the zone of radiation control is analyzed, the influence of social-psychological factors of reproduction of population is considered. On the basis statistic data and results of sociological research the causes of the reduction of insensitivity reproduction behavior of the families in Belarus are revealed. (authors)

  14. Stability and servo-control of the crystal pulling process

    International Nuclear Information System (INIS)

    Johansen, T.H.

    1990-11-01

    The paper analyzes why the crystal pulling process needs servo-control, and how it can be implemented. Special emphasis is put on the fundamental question of inherent stability, and how to interpret the signal from a balance when the weighing method is used for cystal diameter detection. 15 refs., 13 figs

  15. Management control of credit risk in the bank lending process

    NARCIS (Netherlands)

    Scheffer, S.B.

    2004-01-01

    Management control of credit risk in the bank lending processA casestudy to explore improvements from a managerial perspectiveAt the start of this project -back in 1998- new technologies and ideas were emerging among a new generation of financial engineering professionals who have been applying

  16. Serotoninergic regulation of emotional and behavioural control processes.

    NARCIS (Netherlands)

    Cools, R.; Roberts, A.C.; Robbins, T.W.

    2008-01-01

    5-Hydroxytryptamine (5-HT, serotonin) has long been implicated in a wide variety of emotional, cognitive and behavioural control processes. However, its precise contribution is still not well understood. Depletion of 5-HT enhances behavioural and brain responsiveness to punishment or other aversive

  17. Radiosterilization process control in plants using electron accelerators

    International Nuclear Information System (INIS)

    Stuglik, Z.

    1997-01-01

    Electron beam parameters deciding the irradiation dose in radiosterilization plants should be continuously controlled during the process. Dosimetric procedure suitable to irradiated material and dose range should be chosen. The practical advice and directions in this subject have been done. 7 refs

  18. Scheduling algorithms for automatic control systems for technological processes

    Science.gov (United States)

    Chernigovskiy, A. S.; Tsarev, R. Yu; Kapulin, D. V.

    2017-01-01

    Wide use of automatic process control systems and the usage of high-performance systems containing a number of computers (processors) give opportunities for creation of high-quality and fast production that increases competitiveness of an enterprise. Exact and fast calculations, control computation, and processing of the big data arrays - all of this requires the high level of productivity and, at the same time, minimum time of data handling and result receiving. In order to reach the best time, it is necessary not only to use computing resources optimally, but also to design and develop the software so that time gain will be maximal. For this purpose task (jobs or operations), scheduling techniques for the multi-machine/multiprocessor systems are applied. Some of basic task scheduling methods for the multi-machine process control systems are considered in this paper, their advantages and disadvantages come to light, and also some usage considerations, in case of the software for automatic process control systems developing, are made.

  19. Envelopes of Sets of Measures, Tightness, and Markov Control Processes

    International Nuclear Information System (INIS)

    Gonzalez-Hernandez, J.; Hernandez-Lerma, O.

    1999-01-01

    We introduce upper and lower envelopes for sets of measures on an arbitrary topological space, which are then used to give a tightness criterion. These concepts are applied to show the existence of optimal policies for a class of Markov control processes

  20. Quality control procedure for dental x-ray film processing

    International Nuclear Information System (INIS)

    Tingey, D.R.C.

    1983-08-01

    Methods of obtaining the optimum processing from dental films are discussed. A method of quality control of developing conditions for dental x-ray films has been developed. It is relatively easy to maintain and is sufficiently accurate for practical purposes

  1. MIMO Self-Tuning Control of Chemical Process Operation

    DEFF Research Database (Denmark)

    Hallager, L.; Jørgensen, S. B.; Goldschmidt, L.

    1984-01-01

    The problem of selecting a feasible model structure for a MIMO self-tuning controller (MIMOSC) is addressed. The dependency of the necessary structure complexity in relation to the specific process operating point is investigated. Experimental results from a fixed-bed chemical reactor are used...

  2. Statistical Inference at Work: Statistical Process Control as an Example

    Science.gov (United States)

    Bakker, Arthur; Kent, Phillip; Derry, Jan; Noss, Richard; Hoyles, Celia

    2008-01-01

    To characterise statistical inference in the workplace this paper compares a prototypical type of statistical inference at work, statistical process control (SPC), with a type of statistical inference that is better known in educational settings, hypothesis testing. Although there are some similarities between the reasoning structure involved in…

  3. Statistical Process Control in the Practice of Program Evaluation.

    Science.gov (United States)

    Posavac, Emil J.

    1995-01-01

    A technique developed to monitor the quality of manufactured products, statistical process control (SPC), incorporates several features that may prove attractive to evaluators. This paper reviews the history of SPC, suggests how the approach can enrich program evaluation, and illustrates its use in a hospital-based example. (SLD)

  4. Statistical Process Control. Impact and Opportunities for Ohio.

    Science.gov (United States)

    Brown, Harold H.

    The first purpose of this study is to help the reader become aware of the evolution of Statistical Process Control (SPC) as it is being implemented and used in industry today. This is approached through the presentation of a brief historical account of SPC, from its inception through the technological miracle that has occurred in Japan. The…

  5. Statistical Process Control. A Summary. FEU/PICKUP Project Report.

    Science.gov (United States)

    Owen, M.; Clark, I.

    A project was conducted to develop a curriculum and training materials to be used in training industrial operatives in statistical process control (SPC) techniques. During the first phase of the project, questionnaires were sent to 685 companies (215 of which responded) to determine where SPC was being used, what type of SPC firms needed, and how…

  6. An Architectural Style for Closed-loop Process-Control

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak; Eriksen, Ole

    2003-01-01

    This report describes an architectural style for distributed closed-loop process control systems with high performance and hard real-time constraints. The style strikes a good balance between the architectural qualities of performance and modifiability/maintainability that traditionally are often...

  7. An Architectural Style for Closed-loop Process-Control

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak

    This report describes an architectural style for distributed closed-loop process control systems with high performance and hard real-time constraints. The style strikes a good balance between the architectural qualities of performance and modifiability/maintainability that traditionally are often...

  8. Nondestructive control of residual stresses during welding and recharge processes

    International Nuclear Information System (INIS)

    Suarez, J.C.; Fernandez, L.M.; Cruz, C.; Merino, F.; Aragon, B.

    1993-01-01

    In this work, the stress state of material during welding and recharge processes is controlled with the help of Barkhausen effect. The changes, occurred in the longitudinal and transversal stress profile are show during deposition of welding rings. It is proved that the stress state of the base-material depends on the amount of recharge layers, deposited on it

  9. Application of Special Cause Control Charts to Green Sand Process

    Directory of Open Access Journals (Sweden)

    Perzyk M.

    2015-12-01

    Full Text Available Statistical Process Control (SPC based on the well known Shewhart control charts, is widely used in contemporary manufacturing industry, including many foundries. However, the classic SPC methods require that the measured quantities, e.g. process or product parameters, are not auto-correlated, i.e. their current values do not depend on the preceding ones. For the processes which do not obey this assumption the Special Cause Control (SCC charts were proposed, utilizing the residual data obtained from the time-series analysis. In the present paper the results of application of SCC charts to a green sand processing system are presented. The tests, made on real industrial data collected in a big iron foundry, were aimed at the comparison of occurrences of out-of-control signals detected in the original data with those appeared in the residual data. It was found that application of the SCC charts reduces numbers of the signals in almost all cases It is concluded that it can be helpful in avoiding false signals, i.e. resulting from predictable factors.

  10. Statistical process control applied to the liquid-fed ceramic melter process

    International Nuclear Information System (INIS)

    Pulsipher, B.A.; Kuhn, W.L.

    1987-09-01

    In this report, an application of control charts to the apparent feed composition of a Liquid-Fed Ceramic Melter (LFCM) is demonstrated by using results from a simulation of the LFCM system. Usual applications of control charts require the assumption of uncorrelated observations over time. This assumption is violated in the LFCM system because of the heels left in tanks from previous batches. Methods for dealing with this problem have been developed to create control charts for individual batches sent to the feed preparation tank (FPT). These control charts are capable of detecting changes in the process average as well as changes in the process variation. All numbers reported in this document were derived from a simulated demonstration of a plausible LFCM system. In practice, site-specific data must be used as input to a simulation tailored to that site. These data directly affect all variance estimates used to develop control charts. 64 refs., 3 figs., 2 tabs

  11. Statistical process control using optimized neural networks: a case study.

    Science.gov (United States)

    Addeh, Jalil; Ebrahimzadeh, Ata; Azarbad, Milad; Ranaee, Vahid

    2014-09-01

    The most common statistical process control (SPC) tools employed for monitoring process changes are control charts. A control chart demonstrates that the process has altered by generating an out-of-control signal. This study investigates the design of an accurate system for the control chart patterns (CCPs) recognition in two aspects. First, an efficient system is introduced that includes two main modules: feature extraction module and classifier module. In the feature extraction module, a proper set of shape features and statistical feature are proposed as the efficient characteristics of the patterns. In the classifier module, several neural networks, such as multilayer perceptron, probabilistic neural network and radial basis function are investigated. Based on an experimental study, the best classifier is chosen in order to recognize the CCPs. Second, a hybrid heuristic recognition system is introduced based on cuckoo optimization algorithm (COA) algorithm to improve the generalization performance of the classifier. The simulation results show that the proposed algorithm has high recognition accuracy. Copyright © 2013 ISA. Published by Elsevier Ltd. All rights reserved.

  12. Subfemtosecond directional control of chemical processes in molecules

    Science.gov (United States)

    Alnaser, Ali S.; Litvinyuk, Igor V.

    2017-02-01

    Laser pulses with a waveform-controlled electric field and broken inversion symmetry establish the opportunity to achieve directional control of molecular processes on a subfemtosecond timescale. Several techniques could be used to break the inversion symmetry of an electric field. The most common ones include combining a fundamental laser frequency with its second harmonic or with higher -frequency pulses (or pulse trains) as well as using few-cycle pulses with known carrier-envelope phase (CEP). In the case of CEP, control over chemical transformations, typically occurring on a timescale of many femtoseconds, is driven by much faster sub-cycle processes of subfemtosecond to few-femtosecond duration. This is possible because electrons are much lighter than nuclei and fast electron motion is coupled to the much slower nuclear motion. The control originates from populating coherent superpositions of different electronic or vibrational states with relative phases that are dependent on the CEP or phase offset between components of a two-color pulse. In this paper, we review the recent progress made in the directional control over chemical processes, driven by intense few-cycle laser pulses a of waveform-tailored electric field, in different molecules.

  13. [Statistical process control applied to intensity modulated radiotherapy pretreatment controls with portal dosimetry].

    Science.gov (United States)

    Villani, N; Gérard, K; Marchesi, V; Huger, S; François, P; Noël, A

    2010-06-01

    The first purpose of this study was to illustrate the contribution of statistical process control for a better security in intensity modulated radiotherapy (IMRT) treatments. This improvement is possible by controlling the dose delivery process, characterized by pretreatment quality control results. So, it is necessary to put under control portal dosimetry measurements (currently, the ionisation chamber measurements were already monitored by statistical process control thanks to statistical process control tools). The second objective was to state whether it is possible to substitute ionisation chamber with portal dosimetry in order to optimize time devoted to pretreatment quality control. At Alexis-Vautrin center, pretreatment quality controls in IMRT for prostate and head and neck treatments were performed for each beam of each patient. These controls were made with an ionisation chamber, which is the reference detector for the absolute dose measurement, and with portal dosimetry for the verification of dose distribution. Statistical process control is a statistical analysis method, coming from industry, used to control and improve the studied process quality. It uses graphic tools as control maps to follow-up process, warning the operator in case of failure, and quantitative tools to evaluate the process toward its ability to respect guidelines: this is the capability study. The study was performed on 450 head and neck beams and on 100 prostate beams. Control charts, showing drifts, both slow and weak, and also both strong and fast, of mean and standard deviation have been established and have shown special cause introduced (manual shift of the leaf gap of the multileaf collimator). Correlation between dose measured at one point, given with the EPID and the ionisation chamber has been evaluated at more than 97% and disagreement cases between the two measurements were identified. The study allowed to demonstrate the feasibility to reduce the time devoted to

  14. Statistical process control applied to intensity modulated radiotherapy pretreatment controls with portal dosimetry

    International Nuclear Information System (INIS)

    Villani, N.; Noel, A.; Villani, N.; Gerard, K.; Marchesi, V.; Huger, S.; Noel, A.; Francois, P.

    2010-01-01

    Purpose The first purpose of this study was to illustrate the contribution of statistical process control for a better security in intensity modulated radiotherapy (I.M.R.T.) treatments. This improvement is possible by controlling the dose delivery process, characterized by pretreatment quality control results. So, it is necessary to put under control portal dosimetry measurements (currently, the ionisation chamber measurements were already monitored by statistical process control thanks to statistical process control tools). The second objective was to state whether it is possible to substitute ionisation chamber with portal dosimetry in order to optimize time devoted to pretreatment quality control. Patients and methods At Alexis-Vautrin center, pretreatment quality controls in I.M.R.T. for prostate and head and neck treatments were performed for each beam of each patient. These controls were made with an ionisation chamber, which is the reference detector for the absolute dose measurement, and with portal dosimetry for the verification of dose distribution. Statistical process control is a statistical analysis method, coming from industry, used to control and improve the studied process quality. It uses graphic tools as control maps to follow-up process, warning the operator in case of failure, and quantitative tools to evaluate the process toward its ability to respect guidelines: this is the capability study. The study was performed on 450 head and neck beams and on 100 prostate beams. Results Control charts, showing drifts, both slow and weak, and also both strong and fast, of mean and standard deviation have been established and have shown special cause introduced (manual shift of the leaf gap of the multi-leaf collimator). Correlation between dose measured at one point, given with the E.P.I.D. and the ionisation chamber has been evaluated at more than 97% and disagreement cases between the two measurements were identified. Conclusion The study allowed to

  15. Hierarchical optimal control of large-scale nonlinear chemical processes.

    Science.gov (United States)

    Ramezani, Mohammad Hossein; Sadati, Nasser

    2009-01-01

    In this paper, a new approach is presented for optimal control of large-scale chemical processes. In this approach, the chemical process is decomposed into smaller sub-systems at the first level, and a coordinator at the second level, for which a two-level hierarchical control strategy is designed. For this purpose, each sub-system in the first level can be solved separately, by using any conventional optimization algorithm. In the second level, the solutions obtained from the first level are coordinated using a new gradient-type strategy, which is updated by the error of the coordination vector. The proposed algorithm is used to solve the optimal control problem of a complex nonlinear chemical stirred tank reactor (CSTR), where its solution is also compared with the ones obtained using the centralized approach. The simulation results show the efficiency and the capability of the proposed hierarchical approach, in finding the optimal solution, over the centralized method.

  16. Economic model predictive control theory, formulations and chemical process applications

    CERN Document Server

    Ellis, Matthew; Christofides, Panagiotis D

    2017-01-01

    This book presents general methods for the design of economic model predictive control (EMPC) systems for broad classes of nonlinear systems that address key theoretical and practical considerations including recursive feasibility, closed-loop stability, closed-loop performance, and computational efficiency. Specifically, the book proposes: Lyapunov-based EMPC methods for nonlinear systems; two-tier EMPC architectures that are highly computationally efficient; and EMPC schemes handling explicitly uncertainty, time-varying cost functions, time-delays and multiple-time-scale dynamics. The proposed methods employ a variety of tools ranging from nonlinear systems analysis, through Lyapunov-based control techniques to nonlinear dynamic optimization. The applicability and performance of the proposed methods are demonstrated through a number of chemical process examples. The book presents state-of-the-art methods for the design of economic model predictive control systems for chemical processes. In addition to being...

  17. Central auditory processing and migraine: a controlled study.

    Science.gov (United States)

    Agessi, Larissa Mendonça; Villa, Thaís Rodrigues; Dias, Karin Ziliotto; Carvalho, Deusvenir de Souza; Pereira, Liliane Desgualdo

    2014-11-08

    This study aimed to verify and compare central auditory processing (CAP) performance in migraine with and without aura patients and healthy controls. Forty-one volunteers of both genders, aged between 18 and 40 years, diagnosed with migraine with and without aura by the criteria of "The International Classification of Headache Disorders" (ICDH-3 beta) and a control group of the same age range and with no headache history, were included. Gaps-in-noise (GIN), Duration Pattern test (DPT) and Dichotic Digits Test (DDT) tests were used to assess central auditory processing performance. The volunteers were divided into 3 groups: Migraine with aura (11), migraine without aura (15), and control group (15), matched by age and schooling. Subjects with aura and without aura performed significantly worse in GIN test for right ear (p = .006), for left ear (p = .005) and for DPT test (p UNIFESP.

  18. A new instrument for statistical process control of thermoset molding

    International Nuclear Information System (INIS)

    Day, D.R.; Lee, H.L.; Shepard, D.D.; Sheppard, N.F.

    1991-01-01

    The recent development of a rugged ceramic mold mounted dielectric sensor and high speed dielectric instrumentation now enables monitoring and statistical process control of production molding over thousands of runs. In this work special instrumentation and software (ICAM-1000) was utilized that automatically extracts critical point during the molding process including flow point, viscosity minimum gel inflection, and reaction endpoint. In addition, other sensors were incorporated to measure temperature and pressure. The critical point as well as temperature and pressure were then recorded during normal production and then plotted in the form of statistical process control (SPC) charts. Experiments have been carried out in RIM, SMC, and RTM type molding operations. The influence of temperature, pressure chemistry, and other variables has been investigated. In this paper examples of both RIM and SMC are discussed

  19. PLS-based memory control scheme for enhanced process monitoring

    KAUST Repository

    Harrou, Fouzi

    2017-01-20

    Fault detection is important for safe operation of various modern engineering systems. Partial least square (PLS) has been widely used in monitoring highly correlated process variables. Conventional PLS-based methods, nevertheless, often fail to detect incipient faults. In this paper, we develop new PLS-based monitoring chart, combining PLS with multivariate memory control chart, the multivariate exponentially weighted moving average (MEWMA) monitoring chart. The MEWMA are sensitive to incipient faults in the process mean, which significantly improves the performance of PLS methods and widen their applicability in practice. Using simulated distillation column data, we demonstrate that the proposed PLS-based MEWMA control chart is more effective in detecting incipient fault in the mean of the multivariate process variables, and outperform the conventional PLS-based monitoring charts.

  20. Fault tolerant control of multivariable processes using auto-tuning PID controller.

    Science.gov (United States)

    Yu, Ding-Li; Chang, T K; Yu, Ding-Wen

    2005-02-01

    Fault tolerant control of dynamic processes is investigated in this paper using an auto-tuning PID controller. A fault tolerant control scheme is proposed composing an auto-tuning PID controller based on an adaptive neural network model. The model is trained online using the extended Kalman filter (EKF) algorithm to learn system post-fault dynamics. Based on this model, the PID controller adjusts its parameters to compensate the effects of the faults, so that the control performance is recovered from degradation. The auto-tuning algorithm for the PID controller is derived with the Lyapunov method and therefore, the model predicted tracking error is guaranteed to converge asymptotically. The method is applied to a simulated two-input two-output continuous stirred tank reactor (CSTR) with various faults, which demonstrate the applicability of the developed scheme to industrial processes.

  1. A modelling and control structure for product quality control in climate-controlled processing of agro-material

    NARCIS (Netherlands)

    Verdijck, G.J.C.; Straten, van G.

    2002-01-01

    In this paper a modelling and control structure for product quality control is presented for a class of operations that processes agro-material. This class can be characterised as climate-controlled operations, such as storage, transport and drying. The basic model consists of three parts. These are

  2. Adaptive neural network controller for the molten steel level control of strip casting processes

    International Nuclear Information System (INIS)

    Chen, Hung Yi; Huang, Shiuh Jer

    2010-01-01

    The twin-roll strip casting process is a steel-strip production method which combines continuous casting and hot rolling processes. The production line from molten liquid steel to the final steel-strip is shortened and the production cost is reduced significantly as compared to conventional continuous casting. The quality of strip casting process depends on many process parameters, such as molten steel level in the pool, solidification position, and roll gap. Their relationships are complex and the strip casting process has the properties of nonlinear uncertainty and time-varying characteristics. It is difficult to establish an accurate process model for designing a model-based controller to monitor the strip quality. In this paper, a model-free adaptive neural network controller is developed to overcome this problem. The proposed control strategy is based on a neural network structure combined with a sliding-mode control scheme. An adaptive rule is employed to on-line adjust the weights of radial basis functions by using the reaching condition of a specified sliding surface. This surface has the on-line learning ability to respond to the system's nonlinear and time-varying behaviors. Since this model-free controller has a simple control structure and small number of control parameters, it is easy to implement. Simulation results, based on a semi experimental system dynamic model and parameters, are executed to show the control performance of the proposed intelligent controller. In addition, the control performance is compared with that of a traditional Pid controller

  3. Statistical process control charts for monitoring military injuries.

    Science.gov (United States)

    Schuh, Anna; Canham-Chervak, Michelle; Jones, Bruce H

    2017-12-01

    An essential aspect of an injury prevention process is surveillance, which quantifies and documents injury rates in populations of interest and enables monitoring of injury frequencies, rates and trends. To drive progress towards injury reduction goals, additional tools are needed. Statistical process control charts, a methodology that has not been previously applied to Army injury monitoring, capitalise on existing medical surveillance data to provide information to leadership about injury trends necessary for prevention planning and evaluation. Statistical process control Shewhart u-charts were created for 49 US Army installations using quarterly injury medical encounter rates, 2007-2015, for active duty soldiers obtained from the Defense Medical Surveillance System. Injuries were defined according to established military injury surveillance recommendations. Charts display control limits three standard deviations (SDs) above and below an installation-specific historical average rate determined using 28 data points, 2007-2013. Charts are available in Army strategic management dashboards. From 2007 to 2015, Army injury rates ranged from 1254 to 1494 unique injuries per 1000 person-years. Installation injury rates ranged from 610 to 2312 injuries per 1000 person-years. Control charts identified four installations with injury rates exceeding the upper control limits at least once during 2014-2015, rates at three installations exceeded the lower control limit at least once and 42 installations had rates that fluctuated around the historical mean. Control charts can be used to drive progress towards injury reduction goals by indicating statistically significant increases and decreases in injury rates. Future applications to military subpopulations, other health outcome metrics and chart enhancements are suggested. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.

  4. Fuzzy Coordinated PI Controller: Application to the Real-Time Pressure Control Process

    Directory of Open Access Journals (Sweden)

    N. Kanagaraj

    2008-01-01

    Full Text Available This paper presents the real-time implementation of a fuzzy coordinated classical PI control scheme for controlling the pressure in a pilot pressure tank system. The fuzzy system has been designed to track the variation parameters in a feedback loop and tune the classical controller to achieve a better control action for load disturbances and set point changes. The error and process inputs are chosen as the inputs of fuzzy system to tune the conventional PI controller according to the process condition. This online conventional controller tuning technique will reduce the human involvement in controller tuning and increase the operating range of the conventional controller. The proposed control algorithm is experimentally implemented for the real-time pressure control of a pilot air tank system and validated using a high-speed 32-bit ARM7 embedded microcontroller board (ATMEL AT91M55800A. To demonstrate the performance of the fuzzy coordinated PI control scheme, results are compared with a classical PI and PI-type fuzzy control method. It is observed that the proposed controller structure is able to quickly track the parameter variation and perform better in load disturbances and also for set point changes.

  5. The product composition control system at Savannah River: Statistical process control algorithm

    International Nuclear Information System (INIS)

    Brown, K.G.

    1994-01-01

    The Defense Waste Processing Facility (DWPF) at the Savannah River Site (SRS) will be used to immobilize the approximately 130 million liters of high-level nuclear waste currently stored at the site in 51 carbon steel tanks. Waste handling operations separate this waste into highly radioactive insoluble sludge and precipitate and less radioactive water soluble salts. In DWPF, precipitate (PHA) is blended with insoluble sludge and ground glass frit to produce melter feed slurry which is continuously fed to the DWPF melter. The melter produces a molten borosilicate glass which is poured into stainless steel canisters for cooling and, ultimately, shipment to and storage in an geologic repository. Described here is the Product Composition Control System (PCCS) process control algorithm. The PCCS is the amalgam of computer hardware and software intended to ensure that the melt will be processable and that the glass wasteform produced will be acceptable. Within PCCS, the Statistical Process Control (SPC) Algorithm is the means which guides control of the DWPF process. The SPC Algorithm is necessary to control the multivariate DWPF process in the face of uncertainties arising from the process, its feeds, sampling, modeling, and measurement systems. This article describes the functions performed by the SPC Algorithm, characterization of DWPF prior to making product, accounting for prediction uncertainty, accounting for measurement uncertainty, monitoring a SME batch, incorporating process information, and advantages of the algorithm. 9 refs., 6 figs

  6. Microstructural evolution and control in laser material processing

    International Nuclear Information System (INIS)

    Kaul, R.; Nath, A.K.

    2005-01-01

    Laser processing, because of its characteristic features, often gives rise to unique microstructure and properties not obtained with other conventional processes. We present various diverse laser processing case studies involving control of microstructure through judicious selection of processing parameters carried out with indigenously developed high power CO 2 lasers. The first study describes microstructural control during end plug laser welding of PFBR fuel pin, involving crack pone alloy D9 tube and type 316 M stainless steel (SS) plug, through preferential displacement of focused laser beam. Crater and associated cracks were eliminated by suitable laser power ramping. Another case study describes how low heat input characteristics of laser cladding process has been exploited for suppressing dilution in 'Colomony 6' deposits on austenitic SS. The results are in sharp contrast to extensive dilution noticed in Colmony 6 hard faced deposits made by GTAW. A novel laser surface melting (LSM) treatment for type 316 (N) SS weld metal has been developed to generate a sensitization-resistant microstructure which leads to enhanced resistance against intergranular corrosion (IGC). IGC resistance of laser treated surface has been found to be critically dependent on laser processing parameters. Experimental observations have been analyzed with thermal simulation. We have also studied the effect of laser beam spatial intensity profile on the microstructure in LSM. We have developed laser-assisted graded hard facing of austenitic SS substrate with Stellite 6 which, in contrast to direct deposition either by laser or GTAW, produced smooth transition in chemical composition and hardness used to control grain coarsening and martensite formation in type 430 SS weldment. Laser rapid manufacturing (LRM) is emerging as a new rapid and cost effective process for low volume fabrication, esp. of expensive materials. The talk will also present microstructural characteristics of laser

  7. The CEBAF control system

    International Nuclear Information System (INIS)

    Watson, W.A. III.

    1995-01-01

    CEBAF has recently upgraded its accelerator control system to use EPICS, a control system toolkit being developed by a collaboration among laboratories in the US and Europe. The migration to EPICS has taken place during a year of intense commissioning activity, with new and old control systems operating concurrently. Existing CAMAC hardware was preserved by adding a CAMAC serial highway link to VME; newer hardware developments are now primarily in VME. Software is distributed among three tiers of computers: first, workstations and X terminals for operator interfaces and high level applications; second, VME single board computers for distributed access to hardware and for local control processing; third, embedded processors where needed for faster closed loop operation. This system has demonstrated the ability to scale EPICS to controlling thousands of devices, including hundreds of embedded processors, with control distributed among dozens of VME processors executing more than 125,000 EPICS database records. To deal with the large size of the control system, CEBAF has integrated an object oriented database, providing data management capabilities for both low level I/O and high level machine modeling. A new callable interface which is control system independent permits access to live EPICS data, data in other Unix processes, and data contained in the object oriented database

  8. Multilevel flow modelling of process plant for diagnosis and control

    International Nuclear Information System (INIS)

    Lind, M.

    1982-08-01

    The paper describes the multilevel flow modelling methodology which can be used to construct functional models of energy and material processing systems. The models describe mass and energy flow topology on different levels of abstraction and represent the hierarchical functional structure of complex systems. A model of a nuclear power plant (PWR) is presented in the paper for illustration. Due to the consistency of the method, multilevel flow models provide specifications of plant goals and functions and may be used as a basis for design of computer-based support systems for the plant operator. Plant control requirements can be derived from the models and due to independence of the actual controller implementation the method may be used as basic for design of control strategies and for the allocation of control tasks to the computer and the plant operator. (author)

  9. Multilevel Flow Modelling of Process Plant for Diagnosis and Control

    DEFF Research Database (Denmark)

    Lind, Morten

    1982-01-01

    The paper describes the multilevel flow modelling methodology which can be used to construct functional models of energy and material processing systems. The models describe mass and energy flow topology on different levels of abstraction and represent the hierarchical functional structure...... of complex systems. A model of a nuclear power plant (PWR) is presented in the paper for illustration. Due to the consistency of the method, multilevel flow models provide specifications of plant goals and functions and may be used as a basis for design of computer-based support systems for the plant...... operator. Plant control requirements can be derived from the models and due to independence of the actual controller implementation the method may be used as a basis for design of control strategies and for the allocation of control tasks to the computer and the plant operator....

  10. Mechanistic Fermentation Models for Process Design, Monitoring, and Control

    DEFF Research Database (Denmark)

    Mears, Lisa; Stocks, Stuart M.; Albæk, Mads Orla

    2017-01-01

    Mechanistic models require a significant investment of time and resources, but their application to multiple stages of fermentation process development and operation can make this investment highly valuable. This Opinion article discusses how an established fermentation model may be adapted...... for application to different stages of fermentation process development: planning, process design, monitoring, and control. Although a longer development time is required for such modeling methods in comparison to purely data-based model techniques, the wide range of applications makes them a highly valuable tool...... for fermentation research and development. In addition, in a research environment, where collaboration is important, developing mechanistic models provides a platform for knowledge sharing and consolidation of existing process understanding....

  11. Thermal control system. [removing waste heat from industrial process spacecraft

    Science.gov (United States)

    Hewitt, D. R. (Inventor)

    1983-01-01

    The temperature of an exothermic process plant carried aboard an Earth orbiting spacecraft is regulated using a number of curved radiator panels accurately positioned in a circular arrangement to form an open receptacle. A module containing the process is insertable into the receptacle. Heat exchangers having broad exterior surfaces extending axially above the circumference of the module fit within arcuate spacings between adjacent radiator panels. Banks of variable conductance heat pipes partially embedded within and thermally coupled to the radiator panels extend across the spacings and are thermally coupled to broad exterior surfaces of the heat exchangers by flanges. Temperature sensors monitor the temperature of process fluid flowing from the module through the heat exchanges. Thermal conduction between the heat exchangers and the radiator panels is regulated by heating a control fluid within the heat pipes to vary the effective thermal length of the heat pipes in inverse proportion to changes in the temperature of the process fluid.

  12. Systematic Integrated Process Design and Control of Binary Element Reactive Distillation Processes

    DEFF Research Database (Denmark)

    Mansouri, Seyed Soheil; Sales-Cruz, Mauricio; Huusom, Jakob Kjøbsted

    2016-01-01

    In this work, integrated process design and control of reactive distillation processes is considered through a computer-aided framework. First, a set of simple design methods for reactive distillation column that are similar in concept to non-reactive distillation design methods are extended...... to design-control of reactive distillation columns. These methods are based on the element concept where the reacting system of compounds is represented as elements. When only two elements are needed to represent the reacting system of more than two compounds, a binary element system is identified....... It is shown that the same design-control principles that apply to a non-reacting binary system of compounds are also valid for a reactive binary system of elements for distillation columns. Application of this framework shows that designing the reactive distillation process at the maximum driving force...

  13. Model-Based Integrated Process Design and Controller Design of Chemical Processes

    DEFF Research Database (Denmark)

    Abd Hamid, Mohd Kamaruddin Bin

    that is typically formulated as a mathematical programming (optimization with constraints) problem is solved by the so-called reverse approach by decomposing it into four sequential hierarchical sub-problems: (i) pre-analysis, (ii) design analysis, (iii) controller design analysis, and (iv) final selection......This thesis describes the development and application of a new systematic modelbased methodology for performing integrated process design and controller design (IPDC) of chemical processes. The new methodology is simple to apply, easy to visualize and efficient to solve. Here, the IPDC problem...... are ordered according to the defined performance criteria (objective function). The final selected design is then verified through rigorous simulation. In the pre-analysis sub-problem, the concepts of attainable region and driving force are used to locate the optimal process-controller design solution...

  14. Study on the instrumentation and control of the fermentaion process

    Energy Technology Data Exchange (ETDEWEB)

    Nanba, Akira

    1988-03-25

    An attempt was made to create models representing various fermentation processes, and these models were applied to a computer-controlled bread yeast culture plant. First, batch ethanol fermentation processes using yeast were analyzed in the light of reaction rate theory, and various model equations were presented. Using these models, analysis was made at various fermentation temperatures. As a result, it was found that the model equations are applicable to ethanol production and that the reaction rate equation is useful for analyzing high-concentration ethanol fermentation processes. Next, heat involved in ethanol fermentation was measured with a calorimeter. In combination with the above models, automatic montitoring based on microcomputer control, data processing, and display on a screen was attempted, with good results. Using the above findings, the ethanol generation reaction in the bread yeast culture was analyzed qualitatively. Culture control was performed with ethanol growth patterns as signals. The above technique has proved to be applicable to industrial bread yeast production. (17 figs, 3 tabs, 29 refs)

  15. Application of Statistical Process Control (SPC in it´s Quality control

    Directory of Open Access Journals (Sweden)

    Carlos Hernández-Pedrera

    2015-12-01

    Full Text Available The overall objective of this paper is to use the SPC to assess the possibility of improving the process of obtaining a sanitary device. As specific objectives we set out to identify the variables to be analyzed to enter the statistical control of process (SPC, analyze possible errors and variations indicated by the control charts in addition to evaluate and compare the results achieved with the study of SPC before and after monitoring direct in the production line were used sampling methods and laboratory replacement to determine the quality of the finished product, then statistical methods were applied seeking to emphasize the importance and contribution from its application to monitor corrective actions and support processes in production. It was shown that the process is under control because the results were found within established control limits. There is a tendency to be displaced toward one end of the boundary, the distribution exceeds the limits, creating the possibility that under certain conditions the process is out of control, the results also showed that the process being within the limits of quality control is operating far from the optimal conditions. In any of the study situations were obtained products outside the limits of weight and discoloration but defective products were obtained.

  16. Controlled versus automatic processes: which is dominant to safety? The moderating effect of inhibitory control.

    Directory of Open Access Journals (Sweden)

    Yaoshan Xu

    Full Text Available This study explores the precursors of employees' safety behaviors based on a dual-process model, which suggests that human behaviors are determined by both controlled and automatic cognitive processes. Employees' responses to a self-reported survey on safety attitudes capture their controlled cognitive process, while the automatic association concerning safety measured by an Implicit Association Test (IAT reflects employees' automatic cognitive processes about safety. In addition, this study investigates the moderating effects of inhibition on the relationship between self-reported safety attitude and safety behavior, and that between automatic associations towards safety and safety behavior. The results suggest significant main effects of self-reported safety attitude and automatic association on safety behaviors. Further, the interaction between self-reported safety attitude and inhibition and that between automatic association and inhibition each predict unique variances in safety behavior. Specifically, the safety behaviors of employees with lower level of inhibitory control are influenced more by automatic association, whereas those of employees with higher level of inhibitory control are guided more by self-reported safety attitudes. These results suggest that safety behavior is the joint outcome of both controlled and automatic cognitive processes, and the relative importance of these cognitive processes depends on employees' individual differences in inhibitory control. The implications of these findings for theoretical and practical issues are discussed at the end.

  17. Feedback control of chlorine inductively coupled plasma etch processing

    International Nuclear Information System (INIS)

    Lin Chaung; Leou, K.-C.; Shiao, K.-M.

    2005-01-01

    Feedback control has been applied to poly-Si etch processing using a chlorine inductively coupled plasma. Since the positive ion flux and ion energy incident upon the wafer surface are the key factors that influence the etch rate, the ion current and the root mean square (rms) rf voltage on the wafer stage, which are measured using an impedance meter connected to the wafer stage, are adopted as the controlled variables to enhance etch rate. The actuators are two 13.56 MHz rf power generators, which adjust ion density and ion energy, respectively. The results of closed-loop control show that the advantages of feedback control can be achieved. For example, with feedback control, etch rate variation under the transient chamber wall condition is reduced roughly by a factor of 2 as compared to the open-loop case. In addition, the capability of the disturbance rejection was also investigated. For a gas pressure variation of 20%, the largest etch rate variation is about 2.4% with closed-loop control as compared with as large as about 6% variation using open-loop control. Also the effect of ion current and rms rf voltage on etch rate was studied using 2 2 factorial design whose results were used to derive a model equation. The obtained formula was used to adjust the set point of ion current and rf voltage so that the desired etch rate was obtained

  18. Stepless control system for reciprocating compressors: energy savings + process control improvement

    Energy Technology Data Exchange (ETDEWEB)

    Grande, Alvaro; Wenisch, Markus [Hoerbiger Ventilwerke GmbH and Co KG, Wien (Austria); Jacobs, Denis [HOERBIGER do Brasil Industria de Equipamentos, Cajamar, SP (Brazil)

    2012-07-01

    In the past, the capacity of reciprocating compressors was typically controlled by on/off unloaders (step-control) and recycle valves. But due to the fact that the power ratings of new reciprocating compressors for the oil and gas industry increase significantly, advanced control systems are required to reduce power costs and save energy. On top of that, multi-stage compressors are frequently integrated into complex process plants that demand precise control and operational flexibility. There are several solutions for this equation, but maybe the most successful is the use of the reverse flow principle applied to an electronically controlled and hydraulically actuated suction valve unloaders system. (author)

  19. A Draw-In Sensor for Process Control and Optimization

    International Nuclear Information System (INIS)

    Mahayotsanun, Numpon; Cao, Jian; Peshkin, Michael

    2005-01-01

    Sheet metal forming is one of the major processes in manufacturing and is broadly used due to its high degree of design flexibility and low cost. In the sheet metal forming process, draw-in (planar movement of a sheet periphery) frequently occurs and is one of the most dominated indicators on the success of a forming process. Currently, monitoring and controlling draw-in during each stamping operation requires either time-consuming setup or a significant die modification. Most devices have been used only in laboratory settings. Our goal is to design a draw-in sensor providing high sensitivity in monitoring; ease of setup, measurement and controlling; and eventually be implemented in industry. Our design is based on the mutual inductance principle, which we considered physical factors affecting the characteristics of the draw-in sensor. Two different configurations, single-transducer and double-transducer of our draw-in sensors have been designed and tested. The results showed good linearity, especially for the double-transducer case. The output of the draw-in sensor was affected by the type of sheet metal, dimension of the transducer, and the distance between the transducer and the testing sheet metal. It was found that the result was insensitive to the waviness of the sheet metal if sheet thickness was thin. The invention, implementation, and integration of the draw-in sensor will have an enormous impact on revolutionizing the control of stamping process, will provide solid ground for process variation and uncertainty studies, and ultimately will affect the design decision process

  20. Quality control in the process and in the irradiated food

    International Nuclear Information System (INIS)

    Farrar IV, H.

    1997-01-01

    In the irradiation process, absorbed dose is the key parameter that must be controlled. In general, the minimum absorbed dose needed to accomplish a desired effect, such as insect disinfestation or pathogen reduction, is already known from previous research, and is often prescribed by government regulations. The irradiation process is effective, however, only if the food can tolerate this dose without experiencing unwanted changes in flavor or appearance. The dose that food can tolerate often depends on such things as the variety of the fruit or vegetable, where it was grown, the season in which it was harvested and the length of time between harvesting and irradiation. Once the minimum and maximum doses are established, the irradiator operator must make sure that these dose limits are not exceeded. First, a dose mapping using many dosimeters must be undertaken to determine the locations of the minimum and maximum dose in the overall process load. From then on, the process load must always be the same, and, as a key step in the overall process control, dosimeters need to be placed from time to time only at the minimum or maximum locations. The dosimeters must be calibrated and directly trackable to national or international standards, and a fool-proof method of labelling and segregating irradiated from unirradiated product must be used. Radiation sensitive indicators that may help identify irradiated from unirradiated food should not be relied upon, and are not a substitute fro proper dosimetry. (Author)