WorldWideScience

Sample records for automatic data collection systems

  1. Automatic system for crystallographic data collection and analysis

    International Nuclear Information System (INIS)

    Minor, W.; Cymborowski, M.; Otwinowski, Z.

    2002-01-01

    During the last 10 years the rate of the new protein structures determined by X-ray crystallography has risen about tenfold. The use of high flux sources was instrumental in this growth. There are numerous advantages of using synchrotron radiation for protein crystallography: rapid data collection, use of microcrystals and the ability to conduct measurements at wide range of wavelengths. The rate-limiting step is often the ability to analyze and back up a fast stream of data produced by a multi-module CCD detector. A goal of the newly developed HKL-2000 package is to integrate all computational activities that have to be performed during the data collection experiment. The graphical Command Center of HKL-2000 organizes and forwards the data collection parameters to the display, indexing, strategy, simulation, refinement integration, scaling, and merging tasks. Data acquisition can become a part of data processing (or vice versa), which includes indexing, integration, scaling, and even phasing. The increase in inert band with will provide an opportunity to remotely interact with the experimental setup and perform the synchrotron experiment from the home laboratory. (author)

  2. Draft Automatic Data Acquisition System Plan

    International Nuclear Information System (INIS)

    1987-04-01

    This Automatic Data Acquisition System (ADAS) Plan has been prepared in support of the requirement for detailed site characterization of the Deaf Smith County candidate repository site in salt, and describes the data acquisition system which will be used for unattended data collection from the geotechnical instrumentation installed at the site. Section 1.1 discusses the programmatic background to the plan, Section 1.2 presents the scope and purpose of the plan, and the organization of the document is given in Section 1.3. 31 refs., 34 figs., 8 tabs

  3. MAC, A System for Automatically IPR Identification, Collection and Distribution

    Science.gov (United States)

    Serrão, Carlos

    Controlling Intellectual Property Rights (IPR) in the Digital World is a very hard challenge. The facility to create multiple bit-by-bit identical copies from original IPR works creates the opportunities for digital piracy. One of the most affected industries by this fact is the Music Industry. The Music Industry has supported huge losses during the last few years due to this fact. Moreover, this fact is also affecting the way that music rights collecting and distributing societies are operating to assure a correct music IPR identification, collection and distribution. In this article a system for automating this IPR identification, collection and distribution is presented and described. This system makes usage of advanced automatic audio identification system based on audio fingerprinting technology. This paper will present the details of the system and present a use-case scenario where this system is being used.

  4. Can Automatic Classification Help to Increase Accuracy in Data Collection?

    Directory of Open Access Journals (Sweden)

    Frederique Lang

    2016-09-01

    Full Text Available Purpose: The authors aim at testing the performance of a set of machine learning algorithms that could improve the process of data cleaning when building datasets. Design/methodology/approach: The paper is centered on cleaning datasets gathered from publishers and online resources by the use of specific keywords. In this case, we analyzed data from the Web of Science. The accuracy of various forms of automatic classification was tested here in comparison with manual coding in order to determine their usefulness for data collection and cleaning. We assessed the performance of seven supervised classification algorithms (Support Vector Machine (SVM, Scaled Linear Discriminant Analysis, Lasso and elastic-net regularized generalized linear models, Maximum Entropy, Regression Tree, Boosting, and Random Forest and analyzed two properties: accuracy and recall. We assessed not only each algorithm individually, but also their combinations through a voting scheme. We also tested the performance of these algorithms with different sizes of training data. When assessing the performance of different combinations, we used an indicator of coverage to account for the agreement and disagreement on classification between algorithms. Findings: We found that the performance of the algorithms used vary with the size of the sample for training. However, for the classification exercise in this paper the best performing algorithms were SVM and Boosting. The combination of these two algorithms achieved a high agreement on coverage and was highly accurate. This combination performs well with a small training dataset (10%, which may reduce the manual work needed for classification tasks. Research limitations: The dataset gathered has significantly more records related to the topic of interest compared to unrelated topics. This may affect the performance of some algorithms, especially in their identification of unrelated papers. Practical implications: Although the

  5. Automatic Data Collection Design for Neural Networks Detection of ...

    African Journals Online (AJOL)

    Automated data collection is necessary to alleviate problems inherent in data collection for investigation of management frauds. Once we have gathered a realistic data, several methods then exist for proper analysis and detection of anomalous transactions. However, in Nigeria, collecting fraudulent data is relatively difficult ...

  6. automatic data collection design for neural networks detection

    African Journals Online (AJOL)

    Dr Obe

    Automated data collection is necessary to alleviate problems inherent in data collection for investigation ... (iv) Costs the employing organisation assets, ..... velocity (rate of cash flow over a period of time). ... Mining and Knowledge Discovery,.

  7. Roadway system assessment using bluetooth-based automatic vehicle identification travel time data.

    Science.gov (United States)

    2012-12-01

    This monograph is an exposition of several practice-ready methodologies for automatic vehicle identification (AVI) data collection : systems. This includes considerations in the physical setup of the collection system as well as the interpretation of...

  8. Fully automatic characterization and data collection from crystals of biological macromolecules

    International Nuclear Information System (INIS)

    Svensson, Olof; Malbet-Monaco, Stéphanie; Popov, Alexander; Nurizzo, Didier; Bowler, Matthew W.

    2015-01-01

    A fully automatic system has been developed that performs X-ray centring and characterization of, and data collection from, large numbers of cryocooled crystals without human intervention. Considerable effort is dedicated to evaluating macromolecular crystals at synchrotron sources, even for well established and robust systems. Much of this work is repetitive, and the time spent could be better invested in the interpretation of the results. In order to decrease the need for manual intervention in the most repetitive steps of structural biology projects, initial screening and data collection, a fully automatic system has been developed to mount, locate, centre to the optimal diffraction volume, characterize and, if possible, collect data from multiple cryocooled crystals. Using the capabilities of pixel-array detectors, the system is as fast as a human operator, taking an average of 6 min per sample depending on the sample size and the level of characterization required. Using a fast X-ray-based routine, samples are located and centred systematically at the position of highest diffraction signal and important parameters for sample characterization, such as flux, beam size and crystal volume, are automatically taken into account, ensuring the calculation of optimal data-collection strategies. The system is now in operation at the new ESRF beamline MASSIF-1 and has been used by both industrial and academic users for many different sample types, including crystals of less than 20 µm in the smallest dimension. To date, over 8000 samples have been evaluated on MASSIF-1 without any human intervention

  9. Fully automatic characterization and data collection from crystals of biological macromolecules

    Energy Technology Data Exchange (ETDEWEB)

    Svensson, Olof; Malbet-Monaco, Stéphanie; Popov, Alexander; Nurizzo, Didier, E-mail: nurizzo@esrf.fr [European Synchrotron Radiation Facility, 71 Avenue des Martyrs, CS 40220, 38043 Grenoble (France); Bowler, Matthew W., E-mail: nurizzo@esrf.fr [European Molecular Biology Laboratory, Grenoble Outstation, 71 Avenue des Martyrs, CS 90181, 38042 Grenoble (France); Université Grenoble Alpes–EMBL–CNRS, Grenoble Outstation, 71 Avenue des Martyrs, CS 90181, 38042 Grenoble (France); European Synchrotron Radiation Facility, 71 Avenue des Martyrs, CS 40220, 38043 Grenoble (France)

    2015-07-31

    A fully automatic system has been developed that performs X-ray centring and characterization of, and data collection from, large numbers of cryocooled crystals without human intervention. Considerable effort is dedicated to evaluating macromolecular crystals at synchrotron sources, even for well established and robust systems. Much of this work is repetitive, and the time spent could be better invested in the interpretation of the results. In order to decrease the need for manual intervention in the most repetitive steps of structural biology projects, initial screening and data collection, a fully automatic system has been developed to mount, locate, centre to the optimal diffraction volume, characterize and, if possible, collect data from multiple cryocooled crystals. Using the capabilities of pixel-array detectors, the system is as fast as a human operator, taking an average of 6 min per sample depending on the sample size and the level of characterization required. Using a fast X-ray-based routine, samples are located and centred systematically at the position of highest diffraction signal and important parameters for sample characterization, such as flux, beam size and crystal volume, are automatically taken into account, ensuring the calculation of optimal data-collection strategies. The system is now in operation at the new ESRF beamline MASSIF-1 and has been used by both industrial and academic users for many different sample types, including crystals of less than 20 µm in the smallest dimension. To date, over 8000 samples have been evaluated on MASSIF-1 without any human intervention.

  10. Current position on software for the automatic data acquisition system

    International Nuclear Information System (INIS)

    1988-01-01

    This report describes the current concepts for software to control the operation of the Automatic Data Acquisition System (ADAS) proposed for the Deaf Smith County, Texas, Exploratory Shaft Facility (ESF). The purpose of this report is to provide conceptual details of how the ADAS software will execute the data acquisition function, and how the software will make collected information available to the test personnel, the Data Management Group (DMG), and other authorized users. It is not intended that this report describe all of the ADAS functions in exact detail, but the concepts included herein will form the basis for the formal ADAS functional requirements definition document. 5 refs., 14 figs

  11. Fully automatic characterization and data collection from crystals of biological macromolecules.

    Science.gov (United States)

    Svensson, Olof; Malbet-Monaco, Stéphanie; Popov, Alexander; Nurizzo, Didier; Bowler, Matthew W

    2015-08-01

    Considerable effort is dedicated to evaluating macromolecular crystals at synchrotron sources, even for well established and robust systems. Much of this work is repetitive, and the time spent could be better invested in the interpretation of the results. In order to decrease the need for manual intervention in the most repetitive steps of structural biology projects, initial screening and data collection, a fully automatic system has been developed to mount, locate, centre to the optimal diffraction volume, characterize and, if possible, collect data from multiple cryocooled crystals. Using the capabilities of pixel-array detectors, the system is as fast as a human operator, taking an average of 6 min per sample depending on the sample size and the level of characterization required. Using a fast X-ray-based routine, samples are located and centred systematically at the position of highest diffraction signal and important parameters for sample characterization, such as flux, beam size and crystal volume, are automatically taken into account, ensuring the calculation of optimal data-collection strategies. The system is now in operation at the new ESRF beamline MASSIF-1 and has been used by both industrial and academic users for many different sample types, including crystals of less than 20 µm in the smallest dimension. To date, over 8000 samples have been evaluated on MASSIF-1 without any human intervention.

  12. Automatic Traffic Data Collection under Varying Lighting and Temperature Conditions in Multimodal Environments: Thermal versus Visible Spectrum Video-Based Systems

    Directory of Open Access Journals (Sweden)

    Ting Fu

    2017-01-01

    Full Text Available Vision-based monitoring systems using visible spectrum (regular video cameras can complement or substitute conventional sensors and provide rich positional and classification data. Although new camera technologies, including thermal video sensors, may improve the performance of digital video-based sensors, their performance under various conditions has rarely been evaluated at multimodal facilities. The purpose of this research is to integrate existing computer vision methods for automated data collection and evaluate the detection, classification, and speed measurement performance of thermal video sensors under varying lighting and temperature conditions. Thermal and regular video data was collected simultaneously under different conditions across multiple sites. Although the regular video sensor narrowly outperformed the thermal sensor during daytime, the performance of the thermal sensor is significantly better for low visibility and shadow conditions, particularly for pedestrians and cyclists. Retraining the algorithm on thermal data yielded an improvement in the global accuracy of 48%. Thermal speed measurements were consistently more accurate than for the regular video at daytime and nighttime. Thermal video is insensitive to lighting interference and pavement temperature, solves issues associated with visible light cameras for traffic data collection, and offers other benefits such as privacy, insensitivity to glare, storage space, and lower processing requirements.

  13. The protein crystallography beamline BW6 at DORIS - automatic operation and high-throughput data collection

    CERN Document Server

    Blume, H; Bourenkov, G P; Kosciesza, D; Bartunik, H D

    2001-01-01

    The wiggler beamline BW6 at DORIS has been optimized for de-novo solution of protein structures on the basis of MAD phasing. Facilities for automatic data collection, rapid data transfer and storage, and online processing have been developed which provide adequate conditions for high-throughput applications, e.g., in structural genomics.

  14. Crowd-sourced data collection to support automatic classification of building footprint data

    Science.gov (United States)

    Hecht, Robert; Kalla, Matthias; Krüger, Tobias

    2018-05-01

    Human settlements are mainly formed by buildings with their different characteristics and usage. Despite the importance of buildings for the economy and society, complete regional or even national figures of the entire building stock and its spatial distribution are still hardly available. Available digital topographic data sets created by National Mapping Agencies or mapped voluntarily through a crowd via Volunteered Geographic Information (VGI) platforms (e.g. OpenStreetMap) contain building footprint information but often lack additional information on building type, usage, age or number of floors. For this reason, predictive modeling is becoming increasingly important in this context. The capabilities of machine learning allow for the prediction of building types and other building characteristics and thus, the efficient classification and description of the entire building stock of cities and regions. However, such data-driven approaches always require a sufficient amount of ground truth (reference) information for training and validation. The collection of reference data is usually cost-intensive and time-consuming. Experiences from other disciplines have shown that crowdsourcing offers the possibility to support the process of obtaining ground truth data. Therefore, this paper presents the results of an experimental study aiming at assessing the accuracy of non-expert annotations on street view images collected from an internet crowd. The findings provide the basis for a future integration of a crowdsourcing component into the process of land use mapping, particularly the automatic building classification.

  15. Roadway System Assessment Using Bluetooth-Based Automatic Vehicle Identification Travel Time Data

    OpenAIRE

    Day, Christopher M.; Brennan, Thomas M.; Hainen, Alexander M.; Remias, Stephen M.; Bullock, Darcy M.

    2012-01-01

    This monograph is an exposition of several practice-ready methodologies for automatic vehicle identification (AVI) data collection systems. This includes considerations in the physical setup of the collection system as well as the interpretation of the data. An extended discussion is provided, with examples, demonstrating data techniques for converting the raw data into more concise metrics and views. Examples of statistical before-after tests are also provided. A series of case studies were ...

  16. A geological and geophysical data collection system

    Digital Repository Service at National Institute of Oceanography (India)

    Sudhakar, T.; Afzulpurkar, S.

    A geological and geophysical data collection system using a Personal Computer is described below. The system stores data obtained from various survey systems typically installed in a charter vessel and can be used for similar applications on any...

  17. Type-assisted automatic garbage collection for lock-free data structures

    OpenAIRE

    Yang, Albert Mingkun; Wrigstad, Tobias

    2017-01-01

    We introduce Isolde, an automatic garbage collection scheme designed specifically for managing memory in lock-free data structures, such as stacks, lists, maps and queues. Isolde exists as a plug-in memory manager, designed to sit on-top of another memory manager, and use it's allocator and reclaimer (if exists). Isolde treats a lock-free data structure as a logical heap, isolated from the rest of the program. This allows garbage collection outside of Isolde to take place without affecting th...

  18. The FAST-DATA System: Fully Automatic Stochastic Technology for Data Acquisition, Transmission, and Analysis

    International Nuclear Information System (INIS)

    Albrecht, R.W.; Crowe, R.D.; McGuire, J.J.

    1978-01-01

    The potential to automatically collect, classify, and report on stochastic data (signals with random, time-varying components) from power plants has long been discussed by utilities, government, industries, national laboratories and universities. It has become clear to all concerned that such signals often contain information about plant conditions which may provide the basis for increased plant availability through early detection and warning of developing malfunctions. Maintenance can then be scheduled at opportune times. Inopportune failures of major and minor power plant components are a major cause of down-time and detracts significantly from availability of the plant. A complete system to realize automatic stochastic data processing has been conceptually designed. Development of the FAST-DATA system has been initiated through a program of periodic measurements performed on the vibration and loose parts monitoring system of the Trojan reactor (1130-MW(e)PWR) operated by Portland General Electric Company. The development plan for the system consists of a six-step procedure. The initial steps depend on a significant level of human involvement. In the course of development of the system, the routine duties of operators and analysts are gradually replaced by computerized automatic data handling procedures. In the final configuration, the operator and analysts are completely freed of routine chores by logical machinery. The results achieved to date from actual application of the proof-of-principle system are discussed. The early developmental phases have concentrated on system organization and examination of a representative data base. Preliminary results from the signature analysis program using Trojan data indicate that the performance specifications predicted for the FAST-DATA system are achievable in practice. (author)

  19. NLO error propagation exercise data collection system

    International Nuclear Information System (INIS)

    Keisch, B.; Bieber, A.M. Jr.

    1983-01-01

    A combined automated and manual system for data collection is described. The system is suitable for collecting, storing, and retrieving data related to nuclear material control at a bulk processing facility. The system, which was applied to the NLO operated Feed Materials Production Center, was successfully demonstrated for a selected portion of the facility. The instrumentation consisted of off-the-shelf commercial equipment and provided timeliness, convenience, and efficiency in providing information for generating a material balance and performing error propagation on a sound statistical basis

  20. Automatic data acquisition system for a photovoltaic solar plant

    Energy Technology Data Exchange (ETDEWEB)

    Santos, A.; Barrio, C.L.; Guerra, A.G.

    1986-01-01

    An autonomous monitoring system for photovoltaic solar plants is described. The system is able to collect data about the plant's physical and electrical characteristics and also about the environmental conditions. It may present the results on a display, if requested, but its main function is measuring periodically a set of parameters, including several points in the panel I-V characteristics, in an unattended mode. The data are stored on a magnetic tape for later processing on a computer. The system hardware and software are described, as well as their main functions.

  1. Hand held data collection and monitoring system for nuclear facilities

    Science.gov (United States)

    Brayton, D.D.; Scharold, P.G.; Thornton, M.W.; Marquez, D.L.

    1999-01-26

    Apparatus and method is disclosed for a data collection and monitoring system that utilizes a pen based hand held computer unit which has contained therein interaction software that allows the user to review maintenance procedures, collect data, compare data with historical trends and safety limits, and input new information at various collection sites. The system has a means to allow automatic transfer of the collected data to a main computer data base for further review, reporting, and distribution purposes and uploading updated collection and maintenance procedures. The hand held computer has a running to-do list so sample collection and other general tasks, such as housekeeping are automatically scheduled for timely completion. A done list helps users to keep track of all completed tasks. The built-in check list assures that work process will meet the applicable processes and procedures. Users can hand write comments or drawings with an electronic pen that allows the users to directly interface information on the screen. 15 figs.

  2. Hand held data collection and monitoring system for nuclear facilities

    International Nuclear Information System (INIS)

    Brayton, D.D.; Scharold, P.G.; Thornton, M.W.; Marquez, D.L.

    1999-01-01

    Apparatus and method is disclosed for a data collection and monitoring system that utilizes a pen based hand held computer unit which has contained therein interaction software that allows the user to review maintenance procedures, collect data, compare data with historical trends and safety limits, and input new information at various collection sites. The system has a means to allow automatic transfer of the collected data to a main computer data base for further review, reporting, and distribution purposes and uploading updated collection and maintenance procedures. The hand held computer has a running to-do list so sample collection and other general tasks, such as housekeeping are automatically scheduled for timely completion. A done list helps users to keep track of all completed tasks. The built-in check list assures that work process will meet the applicable processes and procedures. Users can hand write comments or drawings with an electronic pen that allows the users to directly interface information on the screen. 15 figs

  3. Affirmative Action Data Collection and Reporting System.

    Science.gov (United States)

    McLean, Marquita; And Others

    A computerized Affirmative Action Data Collection and Reporting System was implemented at the University of Cincinnati to assist in monitoring the progress of the University's Affirmative Action program. The benefits derived from the system were definitely a contributing factor in the University's successful attempt to have its Affirmative Action…

  4. Monitoring the Performance of the Pedestrian Transfer Function of Train Stations Using Automatic Fare Collection Data

    NARCIS (Netherlands)

    Van den Heuvel, J.P.A.; Hoogenraad, J.H.

    2014-01-01

    Over the last years all train stations in The Netherlands have been equipped with automatic fare collection gates and/or validators. All public transport passengers use a smart card to pay their fare. In this paper we present a monitor for the performance of the pedestrian function of train stations

  5. Multichannel display system with automatic sequential output of analog data

    International Nuclear Information System (INIS)

    Bykovskii, Yu.A.; Gruzinov, A.E.; Lagoda, V.B.

    1989-01-01

    The authors describe a device that, with maximum simplicity and autonomy, permits parallel data display from 16 measuring channels with automatic output to the screen of a storage oscilloscope in ∼ 50 μsec. The described device can be used to study the divergence characteristics of the ion component of plasma sources and in optical and x-ray spectroscopy of pulsed processes. Owing to its compactness and autonomy, the device can be located in the immediate vicinity of the detectors (for example, inside a vacuum chamber), which allows the number of vacuum electrical lead-ins and the induction level to be reduced

  6. Automatic generation of water distribution systems based on GIS data.

    Science.gov (United States)

    Sitzenfrei, Robert; Möderl, Michael; Rauch, Wolfgang

    2013-09-01

    In the field of water distribution system (WDS) analysis, case study research is needed for testing or benchmarking optimisation strategies and newly developed software. However, data availability for the investigation of real cases is limited due to time and cost needed for data collection and model setup. We present a new algorithm that addresses this problem by generating WDSs from GIS using population density, housing density and elevation as input data. We show that the resulting WDSs are comparable to actual systems in terms of network properties and hydraulic performance. For example, comparing the pressure heads for an actual and a generated WDS results in pressure head differences of ±4 m or less for 75% of the supply area. Although elements like valves and pumps are not included, the new methodology can provide water distribution systems of varying levels of complexity (e.g., network layouts, connectivity, etc.) to allow testing design/optimisation algorithms on a large number of networks. The new approach can be used to estimate the construction costs of planned WDSs aimed at addressing population growth or at comparisons of different expansion strategies in growth corridors.

  7. Automatic Data Logging and Quality Analysis System for Mobile Devices

    Directory of Open Access Journals (Sweden)

    Yong-Yi Fanjiang

    2017-01-01

    Full Text Available The testing phase of mobile device products includes two important test projects that must be completed before shipment: the field trial and the beta user trial. During the field trial, the product is certified based on its integration and stability with the local operator’s system, and, during the beta user trial, the product is certified by multiple users regarding its daily use, where the goal is to detect and solve early problems. In the traditional approach used to issue returns, testers must log into a web site, fill out a problem form, and then go through a browser or FTP to upload logs; however, this is inconvenient, and problems are reported slowly. Therefore, we propose an “automatic logging analysis system” (ALAS to construct a convenient test environment and, using a record analysis (log parser program, automate the parsing of log files and have questions automatically sent to the database by the system. Finally, the mean time between failures (MTBF is used to establish measurement indicators for the beta user trial.

  8. Fully automated data collection and processing system on macromolecular crystallography beamlines at the PF

    International Nuclear Information System (INIS)

    Yamada, Yusuke; Hiraki, Masahiko; Matsugaki, Naohiro; Chavas, Leonard M.G.; Igarashi, Noriyuki; Wakatsuki, Soichi

    2012-01-01

    Fully automated data collection and processing system has been developed on macromolecular crystallography beamlines at the Photon Factory. In this system, the sample exchange, centering and data collection are sequentially performed for all samples stored in the sample exchange system at a beamline without any manual operations. Data processing of collected data sets is also performed automatically. These results are stored into the database system, and users can monitor the progress and results of automated experiment via a Web browser. (author)

  9. VXIbus data collection system -- A design study

    International Nuclear Information System (INIS)

    Hacker, U.; Richter, B.; Weinert, A.; Arlt, R.; Lewis, W.; Swinhoe, M.

    1995-01-01

    The German support program has sponsored the work to investigate the VXIbus as integration platform for safeguards instrumentation. This paper will cover the analysis of the user requirements for a VXIbus based monitoring system for integrated safeguards -- primarily for reliable unattended in-field collection of large amounts of data. The goal is to develop a suitable system architecture. The design of the system makes use of the VXIbus standard as the selected hardware platform Based upon the requirement analysis and the overriding need for high reliability and robustness, a systematic investigation of different operating system options, as well as development and integration tools will be considered. For the software implementation cycle high and low level programming tools are required. The identification of the constraints for the programming platform and the tool selection will be presented. Both the strategic approach, the rules for analysis and design work as well as the executive components for the support of the implementation and production cycle are given. Here all the conditions for reliable, unattended and integrated safeguards monitoring systems will be addressed. The definition of the basic and advanced design principles are covered. The paper discusses the results of a study on a system produced to demonstrate a high data rate timer/counter application

  10. System for Collecting Biosignal Data from Multiple Patient Monitoring Systems.

    Science.gov (United States)

    Yoon, Dukyong; Lee, Sukhoon; Kim, Tae Young; Ko, JeongGil; Chung, Wou Young; Park, Rae Woong

    2017-10-01

    Biosignal data include important physiological information. For that reason, many devices and systems have been developed, but there has not been enough consideration of how to collect and integrate raw data from multiple systems. To overcome this limitation, we have developed a system for collecting and integrating biosignal data from two patient monitoring systems. We developed an interface to extract biosignal data from Nihon Kohden and Philips monitoring systems. The Nihon Kohden system has a central server for the temporary storage of raw waveform data, which can be requested using the HL7 protocol. However, the Philips system used in our hospital cannot save raw waveform data. Therefore, our system was connected to monitoring devices using the RS232 protocol. After collection, the data were transformed and stored in a unified format. From September 2016 to August 2017, we collected approximately 117 patient-years of waveform data from 1,268 patients in 79 beds of five intensive care units. Because the two systems use the same data storage format, the application software could be run without compatibility issues. Our system collects biosignal data from different systems in a unified format. The data collected by the system can be used to develop algorithms or applications without the need to consider the source of the data.

  11. Operational performance of Swedish grid connected solar power plants. Automatic data collection; Driftuppfoeljning av svenska naetanslutna solcellsanlaeggningar. Automatisering av datainsamling

    Energy Technology Data Exchange (ETDEWEB)

    Hedstroem, Jonas; Svensson, Stefan

    2006-09-15

    A performance database containing all grid-connected PV-systems in Sweden has been in operation since March 2002. The systems in the database are described in detail and energy production is continuously added in the form of monthly values. The energy production and the system descriptions are published on www.elforsk.se/solenergi. In august 2006 31 active systems were present in the database. As result of the Swedish subsidy program this number is expected to increase to over 100 systems in the next years. The new owners of PV-systems are obliged to report the produced electricity to the authorities at least once a year. In this work we have studied different means to simplify the collection of data. Four different methods are defined. 1. The conversion of readings from energy meters made at arbitrary distance in time into monthly values. 2. Methods to handle data obtained with the monitoring systems provided by different inverter manufactures. 3. Methods to acquire data from PV-systems with energy meters reporting to the green certificate system. 4. Commercial GSM/GPRS monitoring systems. The first method is the minimum level required by the authorities. The second and third methods are the use of equipments that are expected to be used by some PV-systems for other reasons. Method 4 gives a possibility to create a fully automatic collection method. The described GPRS-systems are expected to have an initial cost of roughly 4000 SEK and a yearly fee of 200 SEK (1 SEK {approx} 0.14 USD)

  12. A system for classifying wood-using industries and recording statistics for automatic data processing.

    Science.gov (United States)

    E.W. Fobes; R.W. Rowe

    1968-01-01

    A system for classifying wood-using industries and recording pertinent statistics for automatic data processing is described. Forms and coding instructions for recording data of primary processing plants are included.

  13. Smart-card-based automatic meal record system intervention tool for analysis using data mining approach.

    Science.gov (United States)

    Zenitani, Satoko; Nishiuchi, Hiromu; Kiuchi, Takahiro

    2010-04-01

    The Smart-card-based Automatic Meal Record system for company cafeterias (AutoMealRecord system) was recently developed and used to monitor employee eating habits. The system could be a unique nutrition assessment tool for automatically monitoring the meal purchases of all employees, although it only focuses on company cafeterias and has never been validated. Before starting an interventional study, we tested the reliability of the data collected by the system using the data mining approach. The AutoMealRecord data were examined to determine if it could predict current obesity. All data used in this study (n = 899) were collected by a major electric company based in Tokyo, which has been operating the AutoMealRecord system for several years. We analyzed dietary patterns by principal component analysis using data from the system and extracted 5 major dietary patterns: healthy, traditional Japanese, Chinese, Japanese noodles, and pasta. The ability to predict current body mass index (BMI) with dietary preference was assessed with multiple linear regression analyses, and in the current study, BMI was positively correlated with male gender, preference for "Japanese noodles," mean energy intake, protein content, and frequency of body measurement at a body measurement booth in the cafeteria. There was a negative correlation with age, dietary fiber, and lunchtime cafeteria use (R(2) = 0.22). This regression model predicted "would-be obese" participants (BMI >or= 23) with 68.8% accuracy by leave-one-out cross validation. This shows that there was sufficient predictability of BMI based on data from the AutoMealRecord System. We conclude that the AutoMealRecord system is valuable for further consideration as a health care intervention tool. Copyright 2010 Elsevier Inc. All rights reserved.

  14. Automatic Control and Data Acquisition System for Combustion Laboratory Applications.

    Science.gov (United States)

    1982-10-01

    O VPI Access~.ion FCr- 1473 2 UNCLASSIFIED Approved for public release; distribution unlimited JAutomatic Control and Data Acquisition System for...unit. The CPU/ROK board includes a 16 bit microprocessor chip which decodes and executes all in- structions, and controls all data transfers. The 12K...in the limited memory space of 32K of the HP-85 33 ACQDTA’ 1) Controls DevicesCRAIN ,2) Acquires Photodiods Output$ 3) Stores Data o Disc 1

  15. Development of an automatic sample changer and a data acquisition system

    International Nuclear Information System (INIS)

    Bianchini, Ricardo M.; Estevez, Jorge; Vollmer, Alberto E.; Iglicki, Flora A.

    1999-01-01

    An automatic electro-pneumatic sample changer with a rotating sample holder is described. The changer is coupled through an electronic interface with the data acquisition station. The software to automate the system has been designed. (author)

  16. Low-cost automatic activity data recording system

    Directory of Open Access Journals (Sweden)

    Moraes M.F.D.

    1997-01-01

    Full Text Available We describe a low-cost, high quality device capable of monitoring indirect activity by detecting touch-release events on a conducting surface, i.e., the animal's cage cover. In addition to the detecting sensor itself, the system includes an IBM PC interface for prompt data storage. The hardware/software design, while serving for other purposes, is used to record the circadian activity rhythm pattern of rats with time in an automated computerized fashion using minimal cost computer equipment (IBM PC XT. Once the sensor detects a touch-release action of the rat in the upper portion of the cage, the interface sends a command to the PC which records the time (hours-minutes-seconds when the activity occurred. As a result, the computer builds up several files (one per detector/sensor containing a time list of all recorded events. Data can be visualized in terms of actograms, indicating the number of detections per hour, and analyzed by mathematical tools such as Fast Fourier Transform (FFT or cosinor. In order to demonstrate method validation, an experiment was conducted on 8 Wistar rats under 12/12-h light/dark cycle conditions (lights on at 7:00 a.m.. Results show a biological validation of the method since it detected the presence of circadian activity rhythm patterns in the behavior of the rats

  17. Dialog system for automatic data input/output and processing with two BESM-6 computers

    International Nuclear Information System (INIS)

    Belyaev, Y.N.; Gorlov, Y.P.; Makarychev, S.V.; Monakov, A.A.; Shcherbakov, S.A.

    1985-01-01

    This paper presents a system for conducting experiments with fully automatic processing of data from multichannel recorders in the dialog mode. The system acquires data at a rate of 2.5 . 10 3 readings/sec, processes in real time, and outputs digital and graphical material in a multitasking environment

  18. Statistical data processing with automatic system for environmental radiation monitoring

    International Nuclear Information System (INIS)

    Zarkh, V.G.; Ostroglyadov, S.V.

    1986-01-01

    Practice of statistical data processing for radiation monitoring is exemplified, and some results obtained are presented. Experience in practical application of mathematical statistics methods for radiation monitoring data processing allowed to develop a concrete algorithm of statistical processing realized in M-6000 minicomputer. The suggested algorithm by its content is divided into 3 parts: parametrical data processing and hypotheses test, pair and multiple correlation analysis. Statistical processing programms are in a dialogue operation. The above algorithm was used to process observed data over radioactive waste disposal control region. Results of surface waters monitoring processing are presented

  19. Bluetooth data collection system for planning and arterial management.

    Science.gov (United States)

    2014-08-01

    This report presents the results of a research and development project of an implementable portable wireless traffic data collection system. Utilizing Bluetooth wireless technology as a platform, portable battery powered data collection units housed ...

  20. Automatic data acquisition system in CAMAC for spectrometry

    International Nuclear Information System (INIS)

    Szabo, L.; Szalay, S.; Takacz, P.; Pal, A.

    1981-01-01

    A special memory module with twofold access for spectrometric information is described. Via the direct access entry three regimes are realized: plus one, minus one and adress storage. The visual monitoring is carried out by means of an interface containing a buffer memory for the representation of 4 256-channel spectra and the text on the display. The system work is managed by a controller on the basis of a microprocessor

  1. 10 CFR 95.49 - Security of automatic data processing (ADP) systems.

    Science.gov (United States)

    2010-01-01

    ... 10 Energy 2 2010-01-01 2010-01-01 false Security of automatic data processing (ADP) systems. 95.49 Section 95.49 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) FACILITY SECURITY CLEARANCE AND SAFEGUARDING OF NATIONAL SECURITY INFORMATION AND RESTRICTED DATA Control of Information § 95.49 Security of...

  2. Sensor Systems Collect Critical Aerodynamics Data

    Science.gov (United States)

    2010-01-01

    With the support of Small Business Innovation Research (SBIR) contracts with Dryden Flight Research Center, Tao of Systems Integration Inc. developed sensors and other components that will ultimately form a first-of-its-kind, closed-loop system for detecting, measuring, and controlling aerodynamic forces and moments in flight. The Hampton, Virginia-based company commercialized three of the four planned components, which provide sensing solutions for customers such as Boeing, General Electric, and BMW and are used for applications such as improving wind turbine operation and optimizing air flow from air conditioning systems. The completed system may one day enable flexible-wing aircraft with flight capabilities like those of birds.

  3. Automatic data acquisition system of environmental radiation monitor with a personal computer

    International Nuclear Information System (INIS)

    Ohkubo, Tohru; Nakamura, Takashi.

    1984-05-01

    The automatic data acquisition system of environmental radiation monitor was developed in a low price by using a PET personal computer. The count pulses from eight monitors settled at four site boundaries were transmitted to a radiation control room by a signal transmission device and analyzed by the computer via 12 channel scaler and PET-CAMAC Interface for graphic display and printing. (author)

  4. Manual editing of automatically recorded data in an anesthesia information management system.

    Science.gov (United States)

    Wax, David B; Beilin, Yaakov; Hossain, Sabera; Lin, Hung-Mo; Reich, David L

    2008-11-01

    Anesthesia information management systems allow automatic recording of physiologic and anesthetic data. The authors investigated the prevalence of such data modification in an academic medical center. The authors queried their anesthesia information management system database of anesthetics performed in 2006 and tabulated the counts of data points for automatically recorded physiologic and anesthetic parameters as well as the subset of those data that were manually invalidated by clinicians (both with and without alternate values manually appended). Patient, practitioner, data source, and timing characteristics of recorded values were also extracted to determine their associations with editing of various parameters in the anesthesia information management system record. A total of 29,491 cases were analyzed, 19% of which had one or more data points manually invalidated. Among 58 attending anesthesiologists, each invalidated data in a median of 7% of their cases when working as a sole practitioner. A minority of invalidated values were manually appended with alternate values. Pulse rate, blood pressure, and pulse oximetry were the most commonly invalidated parameters. Data invalidation usually resulted in a decrease in parameter variance. Factors independently associated with invalidation included extreme physiologic values, American Society of Anesthesiologists physical status classification, emergency status, timing (phase of the procedure/anesthetic), presence of an intraarterial catheter, resident or certified registered nurse anesthetist involvement, and procedure duration. Editing of physiologic data automatically recorded in an anesthesia information management system is a common practice and results in decreased variability of intraoperative data. Further investigation may clarify the reasons for and consequences of this behavior.

  5. Automatic calibration and signal switching system for the particle beam fusion research data acquisition facility

    Energy Technology Data Exchange (ETDEWEB)

    Boyer, W.B.

    1979-09-01

    This report describes both the hardware and software components of an automatic calibration and signal system (Autocal) for the data acquisition system for the Sandia particle beam fusion research accelerators Hydra, Proto I, and Proto II. The Autocal hardware consists of off-the-shelf commercial equipment. The various hardware components, special modifications and overall system configuration are described. Special software has been developed to support the Autocal hardware. Software operation and maintenance are described.

  6. Automatic calibration and signal switching system for the particle beam fusion research data acquisition facility

    International Nuclear Information System (INIS)

    Boyer, W.B.

    1979-09-01

    This report describes both the hardware and software components of an automatic calibration and signal system (Autocal) for the data acquisition system for the Sandia particle beam fusion research accelerators Hydra, Proto I, and Proto II. The Autocal hardware consists of off-the-shelf commercial equipment. The various hardware components, special modifications and overall system configuration are described. Special software has been developed to support the Autocal hardware. Software operation and maintenance are described

  7. University of Washington Nuclear Physics Data Collection System

    International Nuclear Information System (INIS)

    Green, K.C.

    1981-01-01

    During the past several years, a new data collection system has been developed, replacing a previous system based on an SDS-930 computer. The system is constructed about a PDP 11/60 and an MBD-11 controlled CAMAC crate. The hardware configuration as well as a locally written singles data collection code will be described in some detail. Multiparameter data is taken with an enhanced version of Fermilab Multi. Current capabilities and future plans are discussed

  8. Pilot program for an automated data collection system

    International Nuclear Information System (INIS)

    Burns, R.S.; Johnson, P.S.; Denny, E.C.

    1984-01-01

    This report describes the pilot program of an automated data collection system and presents some of the managerial experiences during its startup. The pilot program demonstrated that improvements can be made in data collection and handling, even when a key hardware item does not meet requirements. 2 figures, 1 table

  9. Mobile In Vivo Infrared Data Collection and Diagnoses Comparison System

    Science.gov (United States)

    Mintz, Frederick W. (Inventor); Moynihan, Philip I. (Inventor); Gunapala, Sarath D. (Inventor)

    2013-01-01

    Described is a mobile in vivo infrared brain scan and analysis system. The system includes a data collection subsystem and a data analysis subsystem. The data collection subsystem is a helmet with a plurality of infrared (IR) thermometer probes. Each of the IR thermometer probes includes an IR photodetector capable of detecting IR radiation generated by evoked potentials within a user's skull. The helmet is formed to collect brain data that is reflective of firing neurons in a mobile subject and transmit the brain data to the data analysis subsystem. The data analysis subsystem is configured to generate and display a three-dimensional image that depicts a location of the firing neurons. The data analysis subsystem is also configured to compare the brain data against a library of brain data to detect an anomaly in the brain data, and notify a user of any detected anomaly in the brain data.

  10. Automatic control systems engineering

    International Nuclear Information System (INIS)

    Shin, Yun Gi

    2004-01-01

    This book gives descriptions of automatic control for electrical electronics, which indicates history of automatic control, Laplace transform, block diagram and signal flow diagram, electrometer, linearization of system, space of situation, state space analysis of electric system, sensor, hydro controlling system, stability, time response of linear dynamic system, conception of root locus, procedure to draw root locus, frequency response, and design of control system.

  11. Online information system for data collection of cattle quality

    Science.gov (United States)

    Sugiharti, E.; Arifudin, R.; Putra, A. T.

    2018-03-01

    Innovation and development of the science and technology which proclaimed by the government through Ristekdikti need to be supported. On the other hand, the Department of Animal Husbandry and Fisheries began introducing the Cattle Card system that contains the identity of each farm animal. Therefore, UNNES especially the Department of Computer Science of FMIPA UNNES, need to give positive contribution in the field of Science and Technology to support the manual system of Cattle Card, through the preparation of prototype of the online information system of data collection of cattle in Semarang regency. The main problem is how to monitor the data of cattle quality through online information system in Semarang regency? The purpose of this research is to produce the prototype of an online information system for data collection of cattle quality in Semarang regency. Main activities: (1) Prepare the flowchart of an online system for data collection of cattle quality. (2) Collecting data to obtain data on identity descriptions of each cattle, owners, mutation records, and health records of livestock cattle. (3) Creation of the prototype of an online information system for data collection of cattle quality in Semarang Regency. The results, (1) had been produced the prototype of an online information system for data collection of cattle in the region of Semarang regency. (2) Socialization of the online information system for cattle quality data collection and exploring input from various related stakeholders. (3) There had been a limited trial of prototypes of the system in Pabelan district in the working area of the Department of Animal Husbandry and Fisheries of Semarang regency and succeeded well.

  12. Implementasi Automatic Packet Reporting System (APRS) Untuk Paket Data Pemantauan dan PengukuranUntuk Paket Data Pemantauan dan Pengukuran

    OpenAIRE

    Arief Goeritno

    2016-01-01

    Telah dilakukan implementasi Automatic Packet Reporting System (APRS) untuk paket data pemantauan dan pengukuran melalui tujuan penelitian, berupa: a) penyetelan program aplikasi pada jaringan APRS dan b) pengukuran terhadap penerimaan data berdasarkan kinerja sensor-sensor. Penyetelan program aplikasi APRS merupakan konfigurasi perangkat lunak untuk APRS yang akan digunakan pada stasiun penerimaan data APRS dengan program aplikasi yang biasa digunakan, yaitu hyperterminal dan UI-View 32. Pen...

  13. Automatic analysis of ultrasonic data

    International Nuclear Information System (INIS)

    Horteur, P.; Colin, J.; Benoist, P.; Bonis, M.; Paradis, L.

    1986-10-01

    This paper describes an automatic and self-contained data processing system, transportable on site, able to perform images such as ''A. Scan'', ''B. Scan'', ... to present very quickly the results of the control. It can be used in the case of pressure vessel inspection [fr

  14. Internal plasma diagnostic with a multichannel magnetic probe system using automatic data acquisition

    International Nuclear Information System (INIS)

    Korten, M.; Carolan, P.G.; Sand, F.; Waelbroeck, F.

    1975-04-01

    A 20-channel magnetic probe system inserted into the plasma is used to measure spatial distributions of poloidal and toroidal magnetic fields in the pulsed toroidal high β-experiment TEE. Plasma parameters, e.g. the β-value, toroidal current density and radial pressure distribution were derived applying static equilibrium theory and can be calculated from the measurements. A data acquisition system used in conjuction with a process computer was operated to obtain the experimental data automatically and to perform the multiple computational tasks. The program system described was built to serve as a first stage of a more common software system applicable for computational data handling for different diagnostics of a plasma physics confinement experiment. (orig.) [de

  15. System of automatic control over data Acquisition and Transmission to IGR NNC RK Data Center

    International Nuclear Information System (INIS)

    Komarov, I.I.; Gordienko, D.D.; Kunakov, A.V.

    2005-01-01

    Automated system for seismic and acoustic data acquisition and transmission in real time was established in Data Center IGR NNC RK, which functions very successively. The system monitors quality and volume of acquired information and also controls the status of the system and communication channels. Statistical data on system operation are accumulated in created database. Information on system status is reflected on the Center Web page. (author)

  16. An airborne meteorological data collection system using satellite relay /ASDAR/

    Science.gov (United States)

    Bagwell, J. W.; Lindow, B. G.

    1978-01-01

    The paper describes the aircraft to satellite data relay (ASDAR) project which processes information collected by the navigation and data systems of widebody jet aircraft which cross data-sparse areas of the tropics and southern hemisphere. The ASDAR system consists of a data acquisition and control unit to acquire, store, and format latitude, longitude, altitude, wind speed, wind direction, and outside air temperature data; a transmitter to relay the formatted data via satellite to the ground; and a clock to time the data sampling and transmission periods.

  17. A Cloud-Based System for Automatic Hazard Monitoring from Sentinel-1 SAR Data

    Science.gov (United States)

    Meyer, F. J.; Arko, S. A.; Hogenson, K.; McAlpin, D. B.; Whitley, M. A.

    2017-12-01

    Despite the all-weather capabilities of Synthetic Aperture Radar (SAR), and its high performance in change detection, the application of SAR for operational hazard monitoring was limited in the past. This has largely been due to high data costs, slow product delivery, and limited temporal sampling associated with legacy SAR systems. Only since the launch of ESA's Sentinel-1 sensors have routinely acquired and free-of-charge SAR data become available, allowing—for the first time—for a meaningful contribution of SAR to disaster monitoring. In this paper, we present recent technical advances of the Sentinel-1-based SAR processing system SARVIEWS, which was originally built to generate hazard products for volcano monitoring centers. We outline the main functionalities of SARVIEWS including its automatic database interface to Sentinel-1 holdings of the Alaska Satellite Facility (ASF), and its set of automatic processing techniques. Subsequently, we present recent system improvements that were added to SARVIEWS and allowed for a vast expansion of its hazard services; specifically: (1) In early 2017, the SARVIEWS system was migrated into the Amazon Cloud, providing access to cloud capabilities such as elastic scaling of compute resources and cloud-based storage; (2) we co-located SARVIEWS with ASF's cloud-based Sentinel-1 archive, enabling the efficient and cost effective processing of large data volumes; (3) we integrated SARVIEWS with ASF's HyP3 system (http://hyp3.asf.alaska.edu/), providing functionality such as subscription creation via API or map interface as well as automatic email notification; (4) we automated the production chains for seismic and volcanic hazards by integrating SARVIEWS with the USGS earthquake notification service (ENS) and the USGS eruption alert system. Email notifications from both services are parsed and subscriptions are automatically created when certain event criteria are met; (5) finally, SARVIEWS-generated hazard products are now

  18. Volunteer-based distributed traffic data collection system

    DEFF Research Database (Denmark)

    Balachandran, Katheepan; Broberg, Jacob Honoré; Revsbech, Kasper

    2010-01-01

    An architecture for a traffic data collection system is proposed, which can collect data without having access to a backbone network. Contrary to other monitoring systems it relies on volunteers to install a program on their own computers, which will capture incoming and outgoing packets, group t...... case performance estimates indicate that this is obtained. Tests conducted by volunteers using an implemented prototype confirm the feasibility of the system......An architecture for a traffic data collection system is proposed, which can collect data without having access to a backbone network. Contrary to other monitoring systems it relies on volunteers to install a program on their own computers, which will capture incoming and outgoing packets, group...... them into flows and send the flow data to a central server. Data can be used for studying and characterising internet traffic and for testing traffic models by regenerating real traffic. The architecture is designed to have efficient and light usage of resources on both client and server sides. Worst...

  19. Automatic Payroll Deposit System.

    Science.gov (United States)

    Davidson, D. B.

    1979-01-01

    The Automatic Payroll Deposit System in Yakima, Washington's Public School District No. 7, directly transmits each employee's salary amount for each pay period to a bank or other financial institution. (Author/MLF)

  20. Automatic Test Systems Aquisition

    National Research Council Canada - National Science Library

    1994-01-01

    We are providing this final memorandum report for your information and use. This report discusses the efforts to achieve commonality in standards among the Military Departments as part of the DoD policy for automatic test systems (ATS...

  1. System for Anonymous Data Collection Based on Group Signature Scheme

    Directory of Open Access Journals (Sweden)

    David Troják

    2016-01-01

    Full Text Available This paper deals with an anonymous data collection in the Internet of Things (IoT. the privacy and anonymity of the data source is important for many IoT applications, such as in agriculture, health, and automotive. the proposed data‑collection system provides anonymity for the data sources by applying a cooperation group scheme. the group scheme also provides a low power consumption. the system is built upon the Tor (The Onion Router anonymous network, which is a part of the Internet darknet. the proposed system was designed for the Android devices on the client side and for Java environment on the server side. We evaluated the anonymous data collection in a real‑use scenario that covers selected data acquisition (e.g. signal strength from smartphones triggered by their geographical location change. the results show that the proposed system provides the sufficient data source anonymity, an effective revocation, a low computational cost and a low overhead.

  2. INfluence of vinasse on water movement in soil, using automatic acquisition and handling data system

    International Nuclear Information System (INIS)

    Nascimento Filho, V.F. do; Barros Ferraz, E.S. de

    1986-01-01

    The vinasse, by-product of ethylic alcohol industry from the sugar cane juice or molasses yeast fermentation, has been incorporated in the soil as fertilizer, due to the its hight organic matter (2-6%), potassium and sulphate (0,1-0,5%) and other nutrient contents. By employing monoenergetic gamma-ray beam attenuation technique (241Am; 59,5 keV; 100 mCi) the influence of vinasse on the water movement in the soil was studied. For this, an automatic acquisition and handling data system was used, based in multichannel analyser, multi-scaling mode operated, coupled to a personal microcomputer and plotter. Despite the small depth studied (6 cm), it was observed that vinasse decreases the water infiltration velocity in the soil. (Author) [pt

  3. A contextual image segmentation system using a priori information for automatic data classification in nuclear physics

    International Nuclear Information System (INIS)

    Benkirane, A.; Auger, G.; Chbihi, A.; Bloyet, D.; Plagnol, E.

    1994-01-01

    This paper presents an original approach to solve an automatic data classification problem by means of image processing techniques. The classification is achieved using image segmentation techniques for extracting the meaningful classes. Two types of information are merged for this purpose: the information contained in experimental images and a priori information derived from underlying physics (and adapted to image segmentation problem). This data fusion is widely used at different stages of the segmentation process. This approach yields interesting results in terms of segmentation performances, even in very noisy cases. Satisfactory classification results are obtained in cases where more ''classical'' automatic data classification methods fail. (authors). 25 refs., 14 figs., 1 append

  4. A contextual image segmentation system using a priori information for automatic data classification in nuclear physics

    Energy Technology Data Exchange (ETDEWEB)

    Benkirane, A; Auger, G; Chbihi, A [Grand Accelerateur National d` Ions Lourds (GANIL), 14 - Caen (France); Bloyet, D [Caen Univ., 14 (France); Plagnol, E [Paris-11 Univ., 91 - Orsay (France). Inst. de Physique Nucleaire

    1994-12-31

    This paper presents an original approach to solve an automatic data classification problem by means of image processing techniques. The classification is achieved using image segmentation techniques for extracting the meaningful classes. Two types of information are merged for this purpose: the information contained in experimental images and a priori information derived from underlying physics (and adapted to image segmentation problem). This data fusion is widely used at different stages of the segmentation process. This approach yields interesting results in terms of segmentation performances, even in very noisy cases. Satisfactory classification results are obtained in cases where more ``classical`` automatic data classification methods fail. (authors). 25 refs., 14 figs., 1 append.

  5. The data collection system for failure/maintenance at the Tritium Systems Test Assembly

    International Nuclear Information System (INIS)

    Casey, M.A.; Gruetzmacher, K.M.; Bartlit, J.R.; Cadwallader, L.C.

    1988-01-01

    A data collection system for obtaining information which can be used to help determine the reliability and vailability of future fusion power plants has been installed at the Los Alamos National Laboratory's Tritium Systems Test Assembly (TSTA). Failure and maintenance data on components of TSTA's tritium systems have been collected since 1984. The focus of the data collection has been TSTA's Tritium Waste Tratment System (TWT), which has maintained high availability since it became operation in 1982. Data collection is still in progress and a total of 291 failure reports are in the data collection system at this time, 47 of which are from the TWT. 6 refs., 2 figs., 2 tabs

  6. The effect of a low-speed automatic brake system estimated from real life data.

    Science.gov (United States)

    Isaksson-Hellman, Irene; Lindman, Magdalena

    2012-01-01

    A substantial part of all traffic accidents involving passenger cars are rear-end collisions and most of them occur at low speed. Auto Brake is a feature that has been launched in several passenger car models during the last few years. City Safety is a technology designed to help the driver mitigate, and in certain situations avoid, rear-end collisions at low speed by automatically braking the vehicle.Studies have been presented that predict promising benefits from these kinds of systems, but few attempts have been made to show the actual effect of Auto Brake. In this study, the effect of City Safety, a standard feature on the Volvo XC60 model, is calculated based on insurance claims data from cars in real traffic crashes in Sweden. The estimated claim frequency of rear-end frontal collisions measured in claims per 1,000 insured vehicle years was 23% lower for the City Safety equipped XC60 model than for other Volvo models without the system.

  7. Automatic, semi-automatic and manual validation of urban drainage data.

    Science.gov (United States)

    Branisavljević, N; Prodanović, D; Pavlović, D

    2010-01-01

    Advances in sensor technology and the possibility of automated long distance data transmission have made continuous measurements the preferable way of monitoring urban drainage processes. Usually, the collected data have to be processed by an expert in order to detect and mark the wrong data, remove them and replace them with interpolated data. In general, the first step in detecting the wrong, anomaly data is called the data quality assessment or data validation. Data validation consists of three parts: data preparation, validation scores generation and scores interpretation. This paper will present the overall framework for the data quality improvement system, suitable for automatic, semi-automatic or manual operation. The first two steps of the validation process are explained in more detail, using several validation methods on the same set of real-case data from the Belgrade sewer system. The final part of the validation process, which is the scores interpretation, needs to be further investigated on the developed system.

  8. Plant operation data collection and database management using NIC system

    International Nuclear Information System (INIS)

    Inase, S.

    1990-01-01

    The Nuclear Information Center (NIC), a division of the Central Research Institute of Electric Power Industry, collects nuclear power plant operation and maintenance information both in Japan and abroad and transmits the information to all domestic utilities so that it can be effectively utilized for safe plant operation and reliability enhancement. The collected information is entered into the database system after being key-worded by NIC. The database system, Nuclear Information database/Communication System (NICS), has been developed by NIC for storage and management of collected information. Objectives of keywords are retrieval and classification by the keyword categories

  9. A preliminary study into performing routine tube output and automatic exposure control quality assurance using radiology information system data

    International Nuclear Information System (INIS)

    Charnock, P.; Jones, R.; Fazakerley, J.; Wilde, R.; Dunn, A. F.

    2011-01-01

    Data are currently being collected from hospital radiology information systems in the North West of the UK for the purposes of both clinical audit and patient dose audit. Could these data also be used to satisfy quality assurance (QA) requirements according to UK guidance? From 2008 to 2009, 731 653 records were submitted from 8 hospitals from the North West England. For automatic exposure control QA, the protocol from Inst. of Physics and Engineering in Medicine (IPEM) report 91 recommends that milli amperes per second can be monitored for repeatability and reproducibility using a suitable phantom, at 70-81 kV. Abdomen AP and chest PA examinations were analysed to find the most common kilo voltage used with these records then used to plot average monthly milli amperes per second with time. IPEM report 91 also recommends that a range of commonly used clinical settings is used to check output reproducibility and repeatability. For each tube, the dose area product values were plotted over time for two most common exposure factor sets. Results show that it is possible to do performance checks of AEC systems; however more work is required to be able to monitor tube output performance. Procedurally, the management system requires work and the benefits to the workflow would need to be demonstrated. (authors)

  10. Detection of clinical mastitis with sensor data from automatic milking systems is improved by using decision-tree induction.

    Science.gov (United States)

    Kamphuis, C; Mollenhorst, H; Heesterbeek, J A P; Hogeveen, H

    2010-08-01

    The objective was to develop and validate a clinical mastitis (CM) detection model by means of decision-tree induction. For farmers milking with an automatic milking system (AMS), it is desirable that the detection model has a high level of sensitivity (Se), especially for more severe cases of CM, at a very high specificity (Sp). In addition, an alert for CM should be generated preferably at the quarter milking (QM) at which the CM infection is visible for the first time. Data were collected from 9 Dutch dairy herds milking automatically during a 2.5-yr period. Data included sensor data (electrical conductivity, color, and yield) at the QM level and visual observations of quarters with CM recorded by the farmers. Visual observations of quarters with CM were combined with sensor data of the most recent automatic milking recorded for that same quarter, within a 24-h time window before the visual assessment time. Sensor data of 3.5 million QM were collected, of which 348 QM were combined with a CM observation. Data were divided into a training set, including two-thirds of all data, and a test set. Cows in the training set were not included in the test set and vice versa. A decision-tree model was trained using only clear examples of healthy (n=24,717) or diseased (n=243) QM. The model was tested on 105 QM with CM and a random sample of 50,000 QM without CM. While keeping the Se at a level comparable to that of models currently used by AMS, the decision-tree model was able to decrease the number of false-positive alerts by more than 50%. At an Sp of 99%, 40% of the CM cases were detected. Sixty-four percent of the severe CM cases were detected and only 12.5% of the CM that were scored as watery milk. The Se increased considerably from 40% to 66.7% when the time window increased from less than 24h before the CM observation, to a time window from 24h before to 24h after the CM observation. Even at very wide time windows, however, it was impossible to reach an Se of 100

  11. Collection and evaluation of salt mixing data with the real time data acquisition system. [LMFBR

    Energy Technology Data Exchange (ETDEWEB)

    Glazer, S.; Chiu, C.; Todreas, N.E.

    1977-09-01

    A minicomputer based real time data acquisition system was designed and built to facilitate data collection during salt mixing tests in mock ups of LMFBR rod bundles. The system represents an expansion of data collection capabilities over previous equipment. It performs steady state and transient monitoring and recording of up to 512 individual electrical resistance probes. Extensive real time software was written to govern all phases of the data collection procedure, including probe definition, probe calibration, salt mixing test data acquisition and storage, and data editing. Offline software was also written to permit data examination and reduction to dimensionless salt concentration maps. Finally, the computer program SUPERENERGY was modified to permit rapid extraction of parameters from dimensionless salt concentration maps. The document describes the computer system, and includes circuit diagrams of all custom built components. It also includes descriptions and listings of all software written, as well as extensive user instructions.

  12. Automatic collection of bovine blood samples | Hale | South African ...

    African Journals Online (AJOL)

    A technique is described which allows automatic collection of jugular venous blood from tethered cows. In this system, blood is pumped continuously from an intravenous cannula which has a double lumen while an anticoagulant is pumped through the second opening. Diluted blood is collected in a fraction collector which ...

  13. Automatic analysis algorithm for radionuclide pulse-height data from beta-gamma coincidence systems

    International Nuclear Information System (INIS)

    Foltz Biegalski, K.M.

    2001-01-01

    There are two acceptable noble gas monitoring measurement modes for Comprehensive Nuclear-Test-Ban-Treaty (CTBT) verification purposes defined in CTBT/PC/II/WG.B/1. These include beta-gamma coincidence and high-resolution gamma-spectrometry. There are at present no commercial, off-the-shelf (COTS) applications for the analysis of β-γ coincidence data. Development of such software is in progress at the Prototype International Data Centre (PIDC) for eventual deployment at the International Data Centre (IDC). Flowcharts detailing the automatic analysis algorithm for β-γ coincidence data to be coded at the PIDC is included. The program is being written in C with Oracle databasing capabilities. (author)

  14. French experience in transient data collection and fatigue monitoring of PWR's nuclear steam supply system

    International Nuclear Information System (INIS)

    Sabaton, M.; Morilhat, P.; Savoldelli, D.; Genette, P.

    1995-10-01

    Electricite de France (EDF), the french national electricity company, is operating 54 standardized pressurizer water reactors. This about 500 reactor-years experience in nuclear stations operation and maintenance area has allowed EDF to develop its own strategy for monitoring of age-related degradations of NPP systems and components relevant for plant safety and reliability. After more than fifteen years of experience in regulatory transient data collection and seven years of successful fatigue monitoring prototypes experimentation, EDF decided to design a new system called SYSFAC (acronym for SYsteme de Surveillance en FAtigue de la Chaudiere) devoted to transient logging and thermal fatigue monitoring of the reactor coolant pressure boundary. The system is fully automatic and directly connected to the on-site data acquisition network without any complementary instrumentation. A functional transient detection module and a mechanical transient detection module are in charge of the general transient data collection. A fatigue monitoring module is aimed towards a precise surveillance of five specific zones particularly sensible to thermal fatigue. After the first step of preliminary studies, the industrial phase of the SYSFAC project is currently going on, with hardware and software tests and implementation. The first SYSFAC system will be delivered to the pilot power plant by the beginning of 1996. The extension to all EDF's nuclear 900 MW is planned after one more year of feedback experience. (authors). 12 refs., 3 figs

  15. Implementasi Automatic Packet Reporting System (APRS Untuk Paket Data Pemantauan dan PengukuranUntuk Paket Data Pemantauan dan Pengukuran

    Directory of Open Access Journals (Sweden)

    Arief Goeritno

    2016-03-01

    Full Text Available Telah dilakukan implementasi Automatic Packet Reporting System (APRS untuk paket data pemantauan dan pengukuran melalui tujuan penelitian, berupa: a penyetelan program aplikasi pada jaringan APRS dan b pengukuran terhadap penerimaan data berdasarkan kinerja sensor-sensor. Penyetelan program aplikasi APRS merupakan konfigurasi perangkat lunak untuk APRS yang akan digunakan pada stasiun penerimaan data APRS dengan program aplikasi yang biasa digunakan, yaitu hyperterminal dan UI-View 32. Pengukuran penerimaan data berdasarkan kinerja sensor yang dilakukan melalui proses perekaman pada stasiun penerimaan data APRS. Kinerja sensor-sensor akan diamati pada stasiun pengiriman dan data hasil pengamatan akan dapat diterima pada stasiun penerimaan secara real time. Program aplikasi berbasis hyperterminal dan UI View 32 telah berhasil melakukan proses handshaking antara Terminal Node Controller (TNC dan komputer, sehingga data telemetri dari stasiun pengiriman paket data dapat diterima di stasiun penerimaan. Data telemetri dapat diamati pada stasiun penerimaan dan dapat diperoleh secara real time dengan format: YB0LRB-11>BEACON,WIDE2-1 [05/18/2014 04:03:49]: : T#010,008,093,004,122,075. Notifikasi YB0LRB-11 merupakan stasiun pengiriman paket data telemetri, kemudian data tersebut akan diterima pada stasiun penerima YD1PRY dengan format: YD1PRY-2>APLPN,ARISS [05/18/2014 04:03:07]: : !06.30.37S/106.48.26E#. Notifikasi tersebut merupakan pengiriman informasi data posisi oleh stasiun YD1PRY untuk inisialisasi pada jaringan APRS. Stasiun YB0LRB-11 ketika mengirim paket data telemetri dengan format: YB0LRB-11>BEACON,WIDE2-1 [05/18/2014 04:00:49]: : T#001,004,035,005,122,075. Paket data dari stasiun YB0LRB yang dipancar ulang atau digipeater dengan format: YB0LRB-11>BEACON,YD1PRY-2,WIDE2* [05/18/2014 04:00:50]: : T#001,004,035,005,122,075. Sensor pengukuran berkinerja relatif stabil, walaupun terdapat nilai simpangan pengukuran sebesar 1 cm atau

  16. Development and Implementation of Production Area of Agricultural Product Data Collection System Based on Embedded System

    Science.gov (United States)

    Xi, Lei; Guo, Wei; Che, Yinchao; Zhang, Hao; Wang, Qiang; Ma, Xinming

    To solve problems in detecting the origin of agricultural products, this paper brings about an embedded data-based terminal, applies middleware thinking, and provides reusable long-range two-way data exchange module between business equipment and data acquisition systems. The system is constructed by data collection node and data center nodes. Data collection nodes taking embedded data terminal NetBoxII as the core, consisting of data acquisition interface layer, controlling information layer and data exchange layer, completing the data reading of different front-end acquisition equipments, and packing the data TCP to realize the data exchange between data center nodes according to the physical link (GPRS / CDMA / Ethernet). Data center node consists of the data exchange layer, the data persistence layer, and the business interface layer, which make the data collecting durable, and provide standardized data for business systems based on mapping relationship of collected data and business data. Relying on public communications networks, application of the system could establish the road of flow of information between the scene of origin certification and management center, and could realize the real-time collection, storage and processing between data of origin certification scene and databases of certification organization, and could achieve needs of long-range detection of agricultural origin.

  17. Computerized techniques for collecting the radioprotection data

    International Nuclear Information System (INIS)

    Cenusa, V.; Valeca, S.; Guta, C.; Talpalariu, C.; Stoica, V.

    2016-01-01

    An important component of a computerized radioprotection system is the module for the collection of the radioprotection data. The data collection can be made automatically from the measurement equipment or manually by the operators after they read the values measured by the mobile devices. Database systems are used for storing the data, they offer higher performances, more efficient data organization, ensure data integrity and controlled access to the data into a multiuser environment. The experimental program for the automatic collection of the remote data transfers periodically, at programmable time intervals, data files from the fixed radiation monitoring stations to a centralized system for radioprotection data. For this is used the File Transfer Protocol (FTP). A Radiation Monitoring Equipment designed and assembled in the Electronics Department of ICN Pitesti was used as a data source for the testing of the experimental programs. (authors)

  18. Automatic digitization of SMA data

    Science.gov (United States)

    Väänänen, Mika; Tanskanen, Eija

    2017-04-01

    In the 1970's and 1980's the Scandinavian Magnetometer Array produced large amounts of excellent data from over 30 stations In Norway, Sweden and Finland. 620 film reels and 20 kilometers of film have been preserved and the longest time series produced in the campaign span almost uninterrupted for five years, but the data has never seen widespread use due to the choice of medium. Film is a difficult medium to digitize efficiently. Previously events of interest were searched for by hand and digitization was done by projecting the film on paper and plotting it by hand. We propose a method of automatically digitizing geomagnetic data stored on film and extracting the numerical values from the digitized data. The automatic digitization process helps in preserving old, valuable data that might otherwise go unused.

  19. Maritime over the Horizon Sensor Integration: High Frequency Surface-Wave-Radar and Automatic Identification System Data Integration Algorithm.

    Science.gov (United States)

    Nikolic, Dejan; Stojkovic, Nikola; Lekic, Nikola

    2018-04-09

    To obtain the complete operational picture of the maritime situation in the Exclusive Economic Zone (EEZ) which lies over the horizon (OTH) requires the integration of data obtained from various sensors. These sensors include: high frequency surface-wave-radar (HFSWR), satellite automatic identification system (SAIS) and land automatic identification system (LAIS). The algorithm proposed in this paper utilizes radar tracks obtained from the network of HFSWRs, which are already processed by a multi-target tracking algorithm and associates SAIS and LAIS data to the corresponding radar tracks, thus forming an integrated data pair. During the integration process, all HFSWR targets in the vicinity of AIS data are evaluated and the one which has the highest matching factor is used for data association. On the other hand, if there is multiple AIS data in the vicinity of a single HFSWR track, the algorithm still makes only one data pair which consists of AIS and HFSWR data with the highest mutual matching factor. During the design and testing, special attention is given to the latency of AIS data, which could be very high in the EEZs of developing countries. The algorithm is designed, implemented and tested in a real working environment. The testing environment is located in the Gulf of Guinea and includes a network of HFSWRs consisting of two HFSWRs, several coastal sites with LAIS receivers and SAIS data provided by provider of SAIS data.

  20. Wireless data collection system for real-time arterial travel time estimates.

    Science.gov (United States)

    2011-03-01

    This project pursued several objectives conducive to the implementation and testing of a Bluetooth (BT) based system to collect travel time data, including the deployment of a BT-based travel time data collection system to perform comprehensive testi...

  1. Automatic computation of radioimmunoassay data

    International Nuclear Information System (INIS)

    Toyota, Takayoshi; Kudo, Mikihiko; Abe, Kanji; Kawamata, Fumiaki; Uehata, Shigeru.

    1975-01-01

    Radioimmunoassay provided dose response curves which showed linearity by the use of logistic transformation (Rodbard). This transformation which was applicable to radioimmunoassay should be useful for the computer processing of insulin and C-peptide assay. In the present studies, standard curves were analysed by testing the fit of analytic functions to radioimmunoassay of insulin and C-peptides. A program for use in combination with the double antibody technique was made by Dr. Kawamata. This approach was evidenced to be useful in order to allow automatic computation of data derived from the double antibody assays of insulin and C-peptides. Automatic corrected calculations of radioimmunoassay data of insulin was found to be satisfactory. (auth.)

  2. [Data collection in anesthesia. Experiences with the inauguration of a new information system].

    Science.gov (United States)

    Zbinden, A M; Rothenbühler, H; Häberli, B

    1997-06-01

    In many institutions information systems are used to process off-line anaesthesia data for invoices, statistical purposes, and quality assurance. Information systems are also increasingly being used to improve process control in order to reduce costs. Most of today's systems were created when information technology and working processes in anaesthesia were very different from those in use today. Thus, many institutions must now replace their computer systems but are probably not aware of how complex this change will be. Modern information systems mostly use client-server architecture and relational data bases. Substituting an old system with a new one is frequently a greater task than designing a system from scratch. This article gives the conclusions drawn from the experience obtained when a large departmental computer system is redesigned in an university hospital. The new system was based on a client-server architecture and was developed by an external company without preceding conceptual analysis. Modules for patient, anaesthesia, surgical, and pain-service data were included. Data were analysed using a separate statistical package (RS/1 from Bolt Beranek), taking advantage of its powerful precompiled procedures. Development and introduction of the new system took much more time and effort than expected despite the use of modern software tools. Introduction of the new program required intensive user training despite the choice of modem graphic screen layouts. Automatic data-reading systems could not be used, as too many faults occurred and the effort for the user was too high. However, after the initial problems were solved the system turned out to be a powerful tool for quality control (both process and outcome quality), billing, and scheduling. The statistical analysis of the data resulted in meaningful and relevant conclusions. Before creating a new information system, the working processes have to be analysed and, if possible, made more efficient; a

  3. Designing a Method for AN Automatic Earthquake Intensities Calculation System Based on Data Mining and On-Line Polls

    Science.gov (United States)

    Liendo Sanchez, A. K.; Rojas, R.

    2013-05-01

    Seismic intensities can be calculated using the Modified Mercalli Intensity (MMI) scale or the European Macroseismic Scale (EMS-98), among others, which are based on a serie of qualitative aspects related to a group of subjective factors that describe human perception, effects on nature or objects and structural damage due to the occurrence of an earthquake. On-line polls allow experts to get an overview of the consequences of an earthquake, without going to the locations affected. However, this could be a hard work if the polls are not properly automated. Taking into account that the answers given to these polls are subjective and there is a number of them that have already been classified for some past earthquakes, it is possible to use data mining techniques in order to automate this process and to obtain preliminary results based on the on-line polls. In order to achieve these goal, a predictive model has been used, using a classifier based on a supervised learning techniques such as decision tree algorithm and a group of polls based on the MMI and EMS-98 scales. It summarized the most important questions of the poll, and recursive divides the instance space corresponding to each question (nodes), while each node splits the space depending on the possible answers. Its implementation was done with Weka, a collection of machine learning algorithms for data mining tasks, using the J48 algorithm which is an implementation of the C4.5 algorithm for decision tree models. By doing this, it was possible to obtain a preliminary model able to identify up to 4 different seismic intensities with 73% correctly classified polls. The error obtained is rather high, therefore, we will update the on-line poll in order to improve the results, based on just one scale, for instance the MMI. Besides, the integration of automatic seismic intensities methodology with a low error probability and a basic georeferencing system, will allow to generate preliminary isoseismal maps

  4. Study to Analyze the Acquisition of Automatic Test Equipment (ATE) Systems. Data Sequence Number A003

    Science.gov (United States)

    1973-12-27

    Systems Test Equipment Comparator, ASTEC ) at NAEC can provide a very accurate Ion a pin by pin basis) match between the UUT and ATE in their data bank...In addition, abbreviated summary data on the ATE is also available to users. ASTEC will also file the UUT data as part of its data bank so that

  5. 15 CFR 911.5 - NOAA Data Collection Systems Use Agreements.

    Science.gov (United States)

    2010-01-01

    ... 15 Commerce and Foreign Trade 3 2010-01-01 2010-01-01 false NOAA Data Collection Systems Use... POLICIES AND PROCEDURES CONCERNING USE OF THE NOAA SPACE-BASED DATA COLLECTION SYSTEMS § 911.5 NOAA Data Collection Systems Use Agreements. (a)(1) In order to use a NOAA DCS, each user must have an agreement with...

  6. Definition of an automatic information retrieval system independent from the data base used

    International Nuclear Information System (INIS)

    Cunha, E.R.

    1983-04-01

    A bibliographic information retrieval system using data stored at the standardized interchange format ISO 2709 or ANSI Z39.2, is specified. A set of comands for interchange format manipulation wich allows the data access at the logical level, achieving the data independence, are used. A data base description language, a storage structure and data base manipulation comands are specified, using retrieval techniques which consider the applications needs. (Author) [pt

  7. Revealing the Linkage Network Dynamic Structures of Chinese Maritime Ports through Automatic Information System Data

    Directory of Open Access Journals (Sweden)

    Hongchu Yu

    2017-10-01

    Full Text Available Marine economic cooperation has emerged as a major theme in this era of globalization; hence, maritime network connectivity and dynamics have attracted more and more attention. Port construction and maritime route improvements increase maritime trade and thus facilitate economic viability and resource sustainability. This paper reveals the regional dimension of inter-port linkage dynamic structure of Chinese maritime ports from a complex multilayer perspective that is meaningful for strategic forecasting and regional long-term economic development planning. In this research, Automatic Information System (AIS-derived traffic flows were used to construct a maritime network and subnetworks based on the geographical locations of ports. The linkage intensity between subnetworks, the linkage tightness within subnetworks, the spatial isolation between high-intensity backbones and tight skeleton networks, and a linkage concentration index for each port were calculated. The ports, in turn, were analyzed based on these network attributes. This study analyzed the external competitiveness and internal cohesion of each subnetwork. The results revealed problems in port management and planning, such as unclear divisions in port operations. More critically, weak complementary relationships between the backbone and skeleton networks among the ports reduce connectivity and must be strengthened. This research contributes to the body of work supporting strategic decision-making for future development.

  8. Source apportionment of PM{sub 10} utilising Automatic Cartridge Collection Unit (ACCU) and TEOM systems in Neath Port Talbot

    Energy Technology Data Exchange (ETDEWEB)

    Hooper, M.J.; Hollingsworth, P. [Neath Port Talbot Borough Council, Neath Port Talbot (United Kingdom); Stone, I.M.; Merefield, J.R. [Exeter Univ., Earth Resources Centre, Exeter (United Kingdom)

    1999-10-01

    The background to this study is the review and assessment of air quality as required by Part IV of the Environment Act 1995. A detailed emissions inventory indicated that there might be a significant problem with PM{sub 10} in the Taibach/Margam area of Port Talbot. This was confirmed by continuous monitoring data from the Automatic Urban Network site in the area at Groeswen hospital. These data, when coupled with wind direction data, enabled the production of pollution roses that showed that the highest average PM{sub 10} levels occurred when the wind was blowing towards the residential area from a large significant local source in the west and south-west, i.e. a steel works. The Council is carrying out a stage 3 assessment of PM{sub 10} in this area with a high probability that an Air Quality Management Action Area (AQMA) will need to be declared. It has therefore become clear, that in order to be able to proceed sensibly to an action plan and attempt to meet Air Quality Objectives, identification of the contribution of PM{sub 10} from specific sources would be required. Initial fingerprinting studies were carried out in partnership with Exeter university and, at an early stage, an Air Quality Management Working Group was formed with the works and the Environment Agency to enable constructive dialogue to take place. (Author)

  9. Automatic remote communication system

    International Nuclear Information System (INIS)

    Yamamoto, Yoichi

    1990-05-01

    The Upgraded RECOVER (Remote Continual Verification) system is a communication system for remote continual verification of security and safeguards status of nuclear material in principal nuclear facilities. The system is composed of a command center and facility sub-systems. A command center is a mini-computer system to process C/S (Containment and Surveillance) status data. Facility sub-systems consists of OSM (On-site Multiplexer), MU (Monitoring Unit) and C/S sensor. The system uses public telephone network for communication between a command center and facility sub-systems, and it encrypts communication data to prevent falsification and wiretapping by unauthorized persons. This system inherits the design principle of RECOVER system that was tested by IAEA before. We upgraded and expanded its capabilities more than those of RECOVER. The development of this system began in 1983, and it finished in 1987. Performance tests of the system were carried out since 1987. It showed a farely good result with some indications which should need further improvements. The Upgraded RECOVER system provides timely information about the status of C/S systems, which could contribute to the reduction of inspection effort and the improvement of cost performance. (author)

  10. 77 FR 56212 - Federal Acquisition Regulation; Information Collection; Use of Data Universal Numbering System...

    Science.gov (United States)

    2012-09-12

    ...; Information Collection; Use of Data Universal Numbering System (DUNS) as Primary Contractor Identification... ``Information Collection 9000-0145, Use of Data Universal Numbering System (DUNS) as Primary Contractor... extension of a previously approved information collection requirement concerning use of the Data Universal...

  11. Automatic generation of data merging program codes.

    OpenAIRE

    Hyensook, Kim; Oussena, Samia; Zhang, Ying; Clark, Tony

    2010-01-01

    Data merging is an essential part of ETL (Extract-Transform-Load) processes to build a data warehouse system. To avoid rewheeling merging techniques, we propose a Data Merging Meta-model (DMM) and its transformation into executable program codes in the manner of model driven engineering. DMM allows defining relationships of different model entities and their merging types in conceptual level. Our formalized transformation described using ATL (ATLAS Transformation Language) enables automatic g...

  12. Making sense of sensor data : detecting clinical mastitis in automatic milking systems

    NARCIS (Netherlands)

    Kamphuis, C.

    2010-01-01

    Farmers milking dairy cows are obliged to exclude milk with abnormal homogeneity or color for human consumption (e.g., Regulation (EC) No 853/2004), where most abnormal milk is caused by clinical mastitis (CM). With automatic milking (AM), farmers are no longer physically present during the milking

  13. Fully automatic CNC machining production system

    Directory of Open Access Journals (Sweden)

    Lee Jeng-Dao

    2017-01-01

    Full Text Available Customized manufacturing is increasing years by years. The consumption habits change has been cause the shorter of product life cycle. Therefore, many countries view industry 4.0 as a target to achieve more efficient and more flexible automated production. To develop an automatic loading and unloading CNC machining system via vision inspection is the first step in industrial upgrading. CNC controller is adopted as the main controller to command to the robot, conveyor, and other equipment in this study. Moreover, machine vision systems are used to detect position of material on the conveyor and the edge of the machining material. In addition, Open CNC and SCADA software will be utilized to make real-time monitor, remote system of control, alarm email notification, and parameters collection. Furthermore, RFID has been added to employee classification and management. The machine handshaking has been successfully proposed to achieve automatic vision detect, edge tracing measurement, machining and system parameters collection for data analysis to accomplish industrial automation system integration with real-time monitor.

  14. Pattern-based Automatic Translation of Structured Power System Data to Functional Models for Decision Support Applications

    DEFF Research Database (Denmark)

    Heussen, Kai; Weckesser, Johannes Tilman Gabriel; Kullmann, Daniel

    2013-01-01

    Improved information and insight for decision support in operations and design are central promises of a smart grid. Well-structured information about the composition of power systems is increasingly becoming available in the domain, e.g. due to standard information models (e.g. CIM or IEC61850......) or otherwise structured databases. More measurements and data do not automatically improve decisions, but there is an opportunity to capitalize on this information for decision support. With suitable reasoning strategies data can be contextualized and decision-relevant events can be promoted and identified....... This paper presents an approach to link available structured power system data directly to a functional representation suitable for diagnostic reasoning. The translation method is applied to test cases also illustrating decision support....

  15. 2012 United States Automatic Identification System Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The 2012 United States Automatic Identification System Database contains vessel traffic data for planning purposes within the U.S. coastal waters. The database is...

  16. 2014 United States Automatic Identification System Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The 2014 United States Automatic Identification System Database contains vessel traffic data for planning purposes within the U.S. coastal waters. The database is...

  17. 2009 United States Automatic Identification System Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The 2009 United States Automatic Identification System Database contains vessel traffic data for planning purposes within the U.S. coastal waters. The database is...

  18. 2010 United States Automatic Identification System Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The 2010 United States Automatic Identification System Database contains vessel traffic data for planning purposes within the U.S. coastal waters. The database is...

  19. 2011 United States Automatic Identification System Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The 2011 United States Automatic Identification System Database contains vessel traffic data for planning purposes within the U.S. coastal waters. The database is...

  20. The collection and analysis of transient test data using the mobile instrumentation data acquisition system (MIDAS)

    International Nuclear Information System (INIS)

    Uncapher, W.L.; Arviso, M.

    1995-01-01

    Packages designed to transport radioactive materials are required to survive exposure to environments defined in Code of Federal Regulations. Cask designers can investigate package designs through structural and thermal testing of full-scale packages, components, or representative models. The acquisition of reliable response data from instrumentation measurement devices is an essential part of this testing activity. Sandia National Laboratories, under the sponsorship of the US Department of Energy (DOE), has developed the Mobile Instrumentation Data Acquisition System (MIDAS) dedicated to the collection and processing of structural and thermal data from regulatory tests

  1. 75 FR 59686 - Proposed Information Collection; Comment Request; NOAA Space-Based Data Collection System (DCS...

    Science.gov (United States)

    2010-09-28

    ...., Washington, DC 20230 (or via the Internet at [email protected] ). FOR FURTHER INFORMATION CONTACT: Requests for... Environmental Satellite (GOES) DCS and the Polar-Orbiting Operational Environmental Satellite (POES) DCS, also... Collection Submittal include Internet, facsimile transmission and postal mailing of paper forms. III. Data...

  2. Feasibility study for adding a demand failure data collection system to the Nuclear Plant Reliability Data System. Final report

    International Nuclear Information System (INIS)

    Tashjian, B.M.; Logan, B.W.

    1983-07-01

    Southwest Research Institute (SwRI) is pleased to submit to Sandia National Laboratories this technical report as fulfillment of Task 5 of the proposal entitled A Feasibility Study for Adding a Duty Cycle Data Collection System to the Nuclear Plant Reliability Data System. The purpose of this report is to summarize the work as delineated in the proposal tasks and to recommend follow-on activities. Technical support for this work was provided by Duke Power Company (Duke), subcontractor to SwRI. The four tasks to be performed in conjunction with the Duty Cycle Data Collection Study (renamed in this report Demand Data Collection) were: define component population and measurable parameters; develop data collection and assessment methodologies; assess the impact on utilities; and assess the impact on NPRDS

  3. A portable wireless data collection system by using optical power supply and photo-communication

    International Nuclear Information System (INIS)

    Nakajima, Toshiro; Shikai, Masahiro; Ikeda, Ikuo; Tochio, Atsushi

    1999-01-01

    For aiming at effective application to annual change management of patrolling inspection data and so forth, a portable wireless measuring and data collection device measurable to vibration, temperature and so forth automatically and for short time under patrolling of inspectors and collectable on sensor signals at many places, to collect field data as electronized data. This device was comprised of a sensor head to mount on an object apparatus to transmit sensor signals and a sensor terminal brought by an inspector and with functions to receive and memory a signal from the sensor head. It had a characteristics capable of wireless data collection using optical power supply and photo-communication where all of power supply to sensor head and transmission and receiving of data were conducted optically. As a result, some characteristics could be realized such as perfect realization of wireless data collection and reduction of maintenance burden without its need on installation of source, signal wire, and so forth, possibility to collect data for short time from distant place, and possibility to conduct high order treatment due to obtaining native waveform signal but no conventional numerical data, and possibility of development on apparatus diagnosis such as detection of abnormal sign and others. (G.K.)

  4. Real time Aanderaa current meter data collection system

    Digital Repository Service at National Institute of Oceanography (India)

    AshokKumar, K.; Diwan, S.G.

    Aanderaa current meters are widely used for recording the current speed and such other 4 parameters by deploying them over extended period of time. Normally data are recorded on magnetic tape and after recovery of current meters, data are read...

  5. System of data collection of muon super-telescope and neutron monitor

    International Nuclear Information System (INIS)

    Klepach, E.; Yanke, V.; Kryakunova, O.; Sarlanis, K.; Souvatsoglou, Zh.; Mavromichalaki, E.

    2005-01-01

    The system of collection of information, integrated with system of selection on concurrences which is easily modified and for collection of the neutron data for the multi directed telescopes and godoscopes is offered. The system of data collection completely is solved at program level on the basis of the super fast processor. Coincidences and decoding of directions of arrival of particles are executed at a program level, and also counters of impulses for necessary number of channels are organized. The system of data collection is executed as the universal external device. Depending on the loaded managing program, this device can be used as: 1) system of telescope data collection, combined with system of selection of double coincidences; or 2) 32-channel system of data collection, for example the neutron monitor; or 3) as the register of the multiple neutrons, generated in the neutron monitor. (author)

  6. AUTOMATIC FREQUENCY CONTROL SYSTEM

    Science.gov (United States)

    Hansen, C.F.; Salisbury, J.D.

    1961-01-10

    A control is described for automatically matching the frequency of a resonant cavity to that of a driving oscillator. The driving oscillator is disconnected from the cavity and a secondary oscillator is actuated in which the cavity is the frequency determining element. A low frequency is mixed with the output of the driving oscillator and the resultant lower and upper sidebands are separately derived. The frequencies of the sidebands are compared with the secondary oscillator frequency. deriving a servo control signal to adjust a tuning element in the cavity and matching the cavity frequency to that of the driving oscillator. The driving oscillator may then be connected to the cavity.

  7. Development of a multiplexer for an automatic data acquisition system for the control and monitoring of microbiological cultures

    Energy Technology Data Exchange (ETDEWEB)

    Morales Rondon, A.; Paredes Puente, J.; Arana Alonso, S.

    2016-07-01

    An automatic data acquisition system has been developed for the control and monitoring of microbiological cultures. Turning an otherwise time-consuming process into a smooth one, by allowing the researcher to set the parameters at the beginning of the experiment and move on into the next task. The development of the hardware and software are key to achieving this system. The mux is custom-made with 22 channels, light weight therefore easy to move around the lab. Furthermore, the software allows the researcher to check the measurements in real-time. It is based on virtual instrumentation software therefore new features can be added easily, thus, the mux is capable of adapting to the scientist necessities. (Author)

  8. Automatically processing physical data from LHD experiments

    Energy Technology Data Exchange (ETDEWEB)

    Emoto, M., E-mail: emoto.masahiko@nifs.ac.jp; Ida, K.; Suzuki, C.; Yoshida, M.; Akiyama, T.; Nakamura, Y.; Sakamoto, R.; Yokoyama, M.; Yoshinuma, M.

    2014-05-15

    Physical data produced by large helical device (LHD) experiments is supplied by the Kaiseki server, and registers more than 200 types of diagnostic data. Dependencies exist amongst the data; i.e., in many cases, the calculation of one data requires other data. Therefore, to obtain unregistered data, one needs to calculate not only the diagnostic data itself but also the dependent data; however, because the data is registered by different scientists, each scientist must separately calculate and register their respective data. To simplify this complicated procedure, we have developed an automatic calculation system called AutoAna. The calculation programs of AutoAna are distributed on a network, and the number of such programs can be easily increased dynamically. Our system is therefore scalable and ready for substantial increases in the size of the target data.

  9. Automatically Identifying and Predicting Unplanned Wind Turbine Stoppages Using SCADA and Alarms System Data: Case Study and Results

    Science.gov (United States)

    Leahy, Kevin; Gallagher, Colm; Bruton, Ken; O'Donovan, Peter; O'Sullivan, Dominic T. J.

    2017-11-01

    Using 10-minute wind turbine SCADA data for fault prediction offers an attractive way of gaining additional prognostic capabilities without needing to invest in extra hardware. To use these data-driven methods effectively, the historical SCADA data must be labelled with the periods when the turbine was in faulty operation as well the sub-system the fault was attributed to. Manually identifying faults using maintenance logs can be effective, but is also highly time consuming and tedious due to the disparate nature of these logs across manufacturers, operators and even individual maintenance events. Turbine alarm systems can help to identify these periods, but the sheer volume of alarms and false positives generated makes analysing them on an individual basis ineffective. In this work, we present a new method for automatically identifying historical stoppages on the turbine using SCADA and alarms data. Each stoppage is associated with either a fault in one of the turbine’s sub-systems, a routine maintenance activity, a grid-related event or a number of other categories. This is then checked against maintenance logs for accuracy and the labelled data fed into a classifier for predicting when these stoppages will occur. Results show that the automated labelling process correctly identifies each type of stoppage, and can be effectively used for SCADA-based prediction of turbine faults.

  10. Automatically controlled training systems

    International Nuclear Information System (INIS)

    Milashenko, A.; Afanasiev, A.

    1990-01-01

    This paper reports that the computer system for NPP personnel training was developed for training centers in the Soviet Union. The system should be considered as the first step in training, taking into account that further steps are to be devoted to part-task and full scope simulator training. The training room consists of 8-12 IBM PC/AT personal computers combined into a network. A trainee accesses the system in a dialor manner. Software enables the instructor to determine the trainee's progress in different subjects of the program. The quality of any trainee preparedness may be evaluated by Knowledge Control operation. Simplified dynamic models are adopted for separate areas of the program. For example, the system of neutron flux monitoring has a dedicated model. Currently, training, requalification and support of professional qualifications of nuclear power plant operators is being emphasized. A significant number of emergency situations during work are occurring due to operator errors. Based on data from September-October 1989, more than half of all unplanned drops in power and stoppages of power plants were due to operator error. As a comparison, problems due to equipment malfunction accounted for no more than a third of the total. The role of personnel, especially of the operators, is significant during normal operations, since energy production costs as well as losses are influenced by the capability of the staff. These facts all point to the importance of quality training of personnel

  11. A Customizable and Expandable Electroencephalography (EEG) Data Collection System

    Science.gov (United States)

    2016-03-01

    PIC18LF46K22T Microcontroller . The ADS1299 is a preferred choice for medical instrumentation devices because of its low-noise PGAs, high-resolution... microcontroller handles the configuration of the ADS1299 and the data communication between the user, or the host computer, and the EEG board. Communication...between the microcontroller and the host computer can be established in 2 ways. The first is serial communication using a Universal Serial Bus to

  12. Effect of external disturbances and data rate on the response of an automatic landing system capable of curved trajectories

    Science.gov (United States)

    Sherman, W. L.

    1975-01-01

    The effects of steady wind, turbulence, data sample rate, and control-actuator natural frequency on the response of a possible automatic landing system were investigated in a nonstatistical study. The results indicate that the system, which interfaces with the microwave landing system, functions well in winds and turbulence as long as the guidance law contains proper compensation for wind. The system response was satisfactory down to five data samples per second, which makes the system compatible with the microwave landing system. No adverse effects were observed when actuator natural frequency was lowered. For limiting cases, those cases where the roll angle goes to zero just as the airplane touches down, the basic method for computing the turn-algorithm gains proved unsatisfactory and unacceptable landings resulted. Revised computation methods gave turn-algorithm gains that resulted in acceptable landings. The gains provided by the new method also improved the touchdown conditions for acceptable landings over those obtained when the gains were determined by the old method.

  13. Summer Student Work Project Report: SCADA Bridge Tool Development Automatically Capturing Data from SCADA to the Maintenance System

    CERN Document Server

    Alhambra-Moron, Alfonso

    2015-01-01

    The main purpose of this report is to summarize the work project I have been doing at CERN during the last 3 months as a Summer Student. My name is Alfonso Alhambra Morón and the 8th of June 2015 I joined the EN-HE-LM team as a summer student supervised by Damien Lafarge in order to collaborate in the automation of the transfer of meter readings from SCADA1 to Infor EAM2, the computerized maintenance management system at CERN. The main objective of my project was to enable the automatic updates of meters in Infor EAM fetching data from SCADA so as to automatize a process which was done manually before and consumed resources in terms of having to consult the meter physically, import this information to Infor EAM by hand and detecting and correcting the errors that can occur when doing all of this manually. This problem is shared by several other teams at CERN apart from the Lift Maintenance team and for this reason the main target I had when developing my solution was flexibility and scalability so as to make...

  14. Video-processing-based system for automated pedestrian data collection and analysis when crossing the street

    Science.gov (United States)

    Mansouri, Nabila; Watelain, Eric; Ben Jemaa, Yousra; Motamed, Cina

    2018-03-01

    Computer-vision techniques for pedestrian detection and tracking have progressed considerably and become widely used in several applications. However, a quick glance at the literature shows a minimal use of these techniques in pedestrian behavior and safety analysis, which might be due to the technical complexities facing the processing of pedestrian videos. To extract pedestrian trajectories from a video automatically, all road users must be detected and tracked during sequences, which is a challenging task, especially in a congested open-outdoor urban space. A multipedestrian tracker based on an interframe-detection-association process was proposed and evaluated. The tracker results are used to implement an automatic tool for pedestrians data collection when crossing the street based on video processing. The variations in the instantaneous speed allowed the detection of the street crossing phases (approach, waiting, and crossing). These were addressed for the first time in the pedestrian road security analysis to illustrate the causal relationship between pedestrian behaviors in the different phases. A comparison with a manual data collection method, by computing the root mean square error and the Pearson correlation coefficient, confirmed that the procedures proposed have significant potential to automate the data collection process.

  15. Automatic River Network Extraction from LIDAR Data

    Science.gov (United States)

    Maderal, E. N.; Valcarcel, N.; Delgado, J.; Sevilla, C.; Ojeda, J. C.

    2016-06-01

    National Geographic Institute of Spain (IGN-ES) has launched a new production system for automatic river network extraction for the Geospatial Reference Information (GRI) within hydrography theme. The goal is to get an accurate and updated river network, automatically extracted as possible. For this, IGN-ES has full LiDAR coverage for the whole Spanish territory with a density of 0.5 points per square meter. To implement this work, it has been validated the technical feasibility, developed a methodology to automate each production phase: hydrological terrain models generation with 2 meter grid size and river network extraction combining hydrographic criteria (topographic network) and hydrological criteria (flow accumulation river network), and finally the production was launched. The key points of this work has been managing a big data environment, more than 160,000 Lidar data files, the infrastructure to store (up to 40 Tb between results and intermediate files), and process; using local virtualization and the Amazon Web Service (AWS), which allowed to obtain this automatic production within 6 months, it also has been important the software stability (TerraScan-TerraSolid, GlobalMapper-Blue Marble , FME-Safe, ArcGIS-Esri) and finally, the human resources managing. The results of this production has been an accurate automatic river network extraction for the whole country with a significant improvement for the altimetric component of the 3D linear vector. This article presents the technical feasibility, the production methodology, the automatic river network extraction production and its advantages over traditional vector extraction systems.

  16. AUTOMATIC RIVER NETWORK EXTRACTION FROM LIDAR DATA

    Directory of Open Access Journals (Sweden)

    E. N. Maderal

    2016-06-01

    Full Text Available National Geographic Institute of Spain (IGN-ES has launched a new production system for automatic river network extraction for the Geospatial Reference Information (GRI within hydrography theme. The goal is to get an accurate and updated river network, automatically extracted as possible. For this, IGN-ES has full LiDAR coverage for the whole Spanish territory with a density of 0.5 points per square meter. To implement this work, it has been validated the technical feasibility, developed a methodology to automate each production phase: hydrological terrain models generation with 2 meter grid size and river network extraction combining hydrographic criteria (topographic network and hydrological criteria (flow accumulation river network, and finally the production was launched. The key points of this work has been managing a big data environment, more than 160,000 Lidar data files, the infrastructure to store (up to 40 Tb between results and intermediate files, and process; using local virtualization and the Amazon Web Service (AWS, which allowed to obtain this automatic production within 6 months, it also has been important the software stability (TerraScan-TerraSolid, GlobalMapper-Blue Marble , FME-Safe, ArcGIS-Esri and finally, the human resources managing. The results of this production has been an accurate automatic river network extraction for the whole country with a significant improvement for the altimetric component of the 3D linear vector. This article presents the technical feasibility, the production methodology, the automatic river network extraction production and its advantages over traditional vector extraction systems.

  17. Curating Virtual Data Collections

    Science.gov (United States)

    Lynnes, Chris; Leon, Amanda; Ramapriyan, Hampapuram; Tsontos, Vardis; Shie, Chung-Lin; Liu, Zhong

    2015-01-01

    NASAs Earth Observing System Data and Information System (EOSDIS) contains a rich set of datasets and related services throughout its many elements. As a result, locating all the EOSDIS data and related resources relevant to particular science theme can be daunting. This is largely because EOSDIS data's organizing principle is affected more by the way they are produced than around the expected end use. Virtual collections oriented around science themes can overcome this by presenting collections of data and related resources that are organized around the user's interest, not around the way the data were produced. Virtual collections consist of annotated web addresses (URLs) that point to data and related resource addresses, thus avoiding the need to copy all of the relevant data to a single place. These URL addresses can be consumed by a variety of clients, ranging from basic URL downloaders (wget, curl) and web browsers to sophisticated data analysis programs such as the Integrated Data Viewer.

  18. On-line measurement with automatic emulsion analysis system and off-line data processing (E531 neutrino experiment)

    International Nuclear Information System (INIS)

    Miyanishi, Motoaki

    1984-01-01

    The automatic emulsion analysis system developed by Nagoya cosmic ray observation group was practically used for the experiment (FNAL-E531) on determining the lifetime of charm particles for the first time in the world, and achieved a great successful result. The system consists of four large precise coordinate-measuring stages capable of conducting simultaneous measurement and multiple (currently four) DOMS (digitized on-line microscope), supported with one mini-computer (ECLIPS S/130). The purpose of E531 experiment was the determination of charm particle lifetime. The experiment was carried out at FNAL, USA, and by the irradiation of wide band ν sub(μ) beam equivalent to 7 x 10 18 of 350 GeV/c protons. The detector was a hybrid system of emulsions and a counter spectrometer. The scan of neutrino reaction, the scan of charm particles, and charm event measurement were analyzed in emulsions, and the on-line programs for-respective analyses were created. Nagoya group has found 726 neutrino reactions in the first run, obtained 37 charm particle candidates, and found 1442 neutrino reactions in the second run, and obtained 56 charm particle candidates. The capability of the automatic emulsion analysis system in terms of the time equired for analysis is in total 3.5 hours per event; 15 minutes for C.S. scan, 15 minutes for coupling to module, 20 minutes for tracing to vertex, 1 hour for neutrino reaction measurement, 10 minutes for offline data processing and 1.5 hours for charm particle scanning. (Wakatsuki, Y.)

  19. The AmeriFlux data activity and data system: an evolving collection of data management techniques, tools, products and services

    Science.gov (United States)

    Boden, T. A.; Krassovski, M.; Yang, B.

    2013-06-01

    The Carbon Dioxide Information Analysis Center (CDIAC) at Oak Ridge National Laboratory (ORNL), USA has provided scientific data management support for the US Department of Energy and international climate change science since 1982. Among the many data archived and available from CDIAC are collections from long-term measurement projects. One current example is the AmeriFlux measurement network. AmeriFlux provides continuous measurements from forests, grasslands, wetlands, and croplands in North, Central, and South America and offers important insight about carbon cycling in terrestrial ecosystems. To successfully manage AmeriFlux data and support climate change research, CDIAC has designed flexible data systems using proven technologies and standards blended with new, evolving technologies and standards. The AmeriFlux data system, comprised primarily of a relational database, a PHP-based data interface and a FTP server, offers a broad suite of AmeriFlux data. The data interface allows users to query the AmeriFlux collection in a variety of ways and then subset, visualize and download the data. From the perspective of data stewardship, on the other hand, this system is designed for CDIAC to easily control database content, automate data movement, track data provenance, manage metadata content, and handle frequent additions and corrections. CDIAC and researchers in the flux community developed data submission guidelines to enhance the AmeriFlux data collection, enable automated data processing, and promote standardization across regional networks. Both continuous flux and meteorological data and irregular biological data collected at AmeriFlux sites are carefully scrutinized by CDIAC using established quality-control algorithms before the data are ingested into the AmeriFlux data system. Other tasks at CDIAC include reformatting and standardizing the diverse and heterogeneous datasets received from individual sites into a uniform and consistent network database

  20. The AmeriFlux data activity and data system: an evolving collection of data management techniques, tools, products and services

    Directory of Open Access Journals (Sweden)

    T. A. Boden

    2013-06-01

    Full Text Available The Carbon Dioxide Information Analysis Center (CDIAC at Oak Ridge National Laboratory (ORNL, USA has provided scientific data management support for the US Department of Energy and international climate change science since 1982. Among the many data archived and available from CDIAC are collections from long-term measurement projects. One current example is the AmeriFlux measurement network. AmeriFlux provides continuous measurements from forests, grasslands, wetlands, and croplands in North, Central, and South America and offers important insight about carbon cycling in terrestrial ecosystems. To successfully manage AmeriFlux data and support climate change research, CDIAC has designed flexible data systems using proven technologies and standards blended with new, evolving technologies and standards. The AmeriFlux data system, comprised primarily of a relational database, a PHP-based data interface and a FTP server, offers a broad suite of AmeriFlux data. The data interface allows users to query the AmeriFlux collection in a variety of ways and then subset, visualize and download the data. From the perspective of data stewardship, on the other hand, this system is designed for CDIAC to easily control database content, automate data movement, track data provenance, manage metadata content, and handle frequent additions and corrections. CDIAC and researchers in the flux community developed data submission guidelines to enhance the AmeriFlux data collection, enable automated data processing, and promote standardization across regional networks. Both continuous flux and meteorological data and irregular biological data collected at AmeriFlux sites are carefully scrutinized by CDIAC using established quality-control algorithms before the data are ingested into the AmeriFlux data system. Other tasks at CDIAC include reformatting and standardizing the diverse and heterogeneous datasets received from individual sites into a uniform and consistent

  1. Automatic control system at the ''Loviisa'' NPP

    International Nuclear Information System (INIS)

    Kukhtevich, I.V.; Mal'tsev, B.K.; Sergievskaya, E.N.

    1980-01-01

    Automatic control system of the Loviisa-1 NPP (Finland) is described. According to operation conditions of Finland power system the Loviisa-1 NPP must operate in the mode of week and day control of loading schedule and participate in current control of power system frequency and capacity. With provision for these requirements NPP is equipped with the all-regime system for automatic control functioning during reactor start-up, shut-down, in normal and transient regimes and in emergency situations. The automatic control system includes: a data subsystem, an automatic control subsystem, a discrete control subsystem including remote, a subsystem for reactor control and protection and overall station system of protections: control and dosimetry inside the reactor. Structures of a data-computer complex, discrete control subsystems, reactor control and protection systems, neutron flux control system, inside-reactor control system, station protection system and system for control of fuel element tightness are presented in short. Two-year experience of the NPP operation confirmed advisability of the chosen volume of automatization. The Loviisa-1 NPP operates successfully in the mode of the week and day control of supervisor schedule and current control of frequency (short-term control)

  2. NOSS/ALDCS analysis and system requirements definition. [national oceanic satellite system data collection

    Science.gov (United States)

    Reed, D. L.; Wallace, R. G.

    1981-01-01

    The results of system analyses and implementation studies of an advanced location and data collection system (ALDCS) , proposed for inclusion on the National Oceanic Satellite System (NOSS) spacecraft are reported. The system applies Doppler processing and radiofrequency interferometer position location technqiues both alone and in combination. Aspects analyzed include: the constraints imposed by random access to the system by platforms, the RF link parameters, geometric concepts of position and velocity estimation by the two techniques considered, and the effects of electrical measurement errors, spacecraft attitude errors, and geometric parameters on estimation accuracy. Hardware techniques and trade-offs for interferometric phase measurement, ambiguity resolution and calibration are considered. A combined Doppler-interferometer ALDCS intended to fulfill the NOSS data validation and oceanic research support mission is also described.

  3. Mobile robot teleoperation system for plant inspection based on collecting and utilizing environment data

    International Nuclear Information System (INIS)

    Kawabata, Kuniaki; Watanabe, Nobuyasu; Asama, Hajime; Kita, Nobuyuki; Yang, Hai-quan

    2004-01-01

    This paper describes about development of a mobile robot teleoperation system for plant inspection. In our system, the robot is an agent for collecting the environment data and is also teleoperated by the operator utilizing such accumulated environment data which is displayed on the operation interface. The robot equips many sensors for detecting the state of the robot and the environment. Such redundant sensory system can be also utilized to collect the working environment data on-site while the robot is patrolling. Here, proposed system introduces the framework of collecting and utilizing environment data for adaptive plant inspection using the teleoperated robot. A view simulator is primarily aiming to facilitate evaluation of the visual sensors and algorithms and is also extended as the Environment Server, which is the core technology of the digital maintenance field for the plant inspection. In order to construct detailed seamless digital maintenance field mobile robotic technology is utilized to supply environment data to the server. The sensory system on the robot collect the environment data on-site and such collected data is uploaded to the Environment Server for compiling accurate digital environment data base. The robot operator also can utilize accumulated environment data by referring to the Environment Server. In this paper, we explain the concept of our teleoperation system based on collecting and utilizing environment data. Using developed system, inspection patrol experiments were attempted in the plant mock-up. Experimental results are shown by using an omnidirectional mobile robot with sensory system and the Environment Server. (author)

  4. Designing automatic resupply systems.

    Science.gov (United States)

    Harding, M L

    1999-02-01

    This article outlines the process for designing and implementing autoresupply systems. The planning process includes determination of goals and appropriate participation. Different types of autoresupply mechanisms include kanban, breadman, consignment, systems contracts, and direct shipping from an MRP schedule.

  5. An automatic measuring system of internal friction at low frequency

    International Nuclear Information System (INIS)

    Iwasaki, K.

    1979-01-01

    An inverted torsion pendulum is automatized by means of Tectanel electronic system. Internal friction and the period of vibration are measured fully automatically as a function of temperature and the data obtained are analysed with a computer. (Author) [pt

  6. Automatic multi-modal intelligent seizure acquisition (MISA) system for detection of motor seizures from electromyographic data and motion data

    DEFF Research Database (Denmark)

    Conradsen, Isa; Beniczky, Sándor; Wolf, Peter

    2012-01-01

    measures of reconstructed sub-bands from the discrete wavelet transformation (DWT) and the wavelet packet transformation (WPT). Based on the extracted features all data segments were classified using a support vector machine (SVM) algorithm as simulated seizure or normal activity. A case study...... of the seizure from the patient showed that the simulated seizures were visually similar to the epileptic one. The multi-modal intelligent seizure acquisition (MISA) system showed high sensitivity, short detection latency and low false detection rate. The results showed superiority of the multi- modal detection...... system compared to the uni-modal one. The presented system has a promising potential for seizure detection based on multi-modal data....

  7. Using sensor data patterns from an automatic milking system to develop predictive variables for classifying clinical mastitis and abnormal milk

    NARCIS (Netherlands)

    Kamphuis, A.; Pietersma, D.; Tol, van der R.; Wiedermann, M.; Hogeveen, H.

    2008-01-01

    Dairy farmers using automatic milking are able to manage mastitis successfully with the help of mastitis attention lists. These attention lists are generated with mastitis detection models that make use of sensor data obtained throughout each quarter milking. The models tend to be limited to using

  8. Automatic bagout system

    International Nuclear Information System (INIS)

    Whitaker, R.M.

    1986-11-01

    Nuclear material entrained wastes are generated at the Plutonium Facility at Los Alamos National Laboratory. These wastes are removed from the glove box lines using the bagout method. This is a manual operation performed by technicians. An automated system is being developed to relieve the technicians from this task. The system will reduce the amount of accumulated radiation exposure to the worker. The primary components of the system consist of a six degree of freedom robot, a bag sealing device, and a small gantry robot. 1 ref., 5 figs

  9. Automatic bagout system

    International Nuclear Information System (INIS)

    Whitaker, R.M.

    1987-01-01

    Nuclear material entrained wastes are generated at the plutonium facility at Los Alamos National Laboratory. These wastes are removed from the glove box lines using the bagout method. This is a manual operation performed by technicians. An automated system is being developed to relieve the technicians from this task. The system will reduce the amount of accumulated radiation exposure to the worker. The primary components of the system consist of a six degree of freedom robot, a nag sealing device, and a small gantry robot

  10. Automatic classification of municipal call data for quantitative urban drainage system analysis

    NARCIS (Netherlands)

    Ten Veldhuis, J.A.E.; Harder, R.C.; Loog, M.

    2010-01-01

    Flooding in urban areas can be caused by heavy rainfall, improper planning or component failures. Quantification of these various causes to urban flood probability supports prioritisation of flood risk reduction measures. In many cases, a lack of data on flooding incidents impedes quantification of

  11. 77 FR 39985 - Information Collection; Forest Industries and Residential Fuelwood and Post Data Collection Systems

    Science.gov (United States)

    2012-07-06

    ... personnel use three questionnaires, which are collected by personal mill visits or phone calls, or which respondents return in self-addressed, postage pre-paid envelopes, or by email. Pulpwood Received Questionnaire: Forest Service personnel use this questionnaire to collect and evaluate information from pulp and...

  12. Data Collection Guidelines for Consistent Evaluation of Data from Verification and Monitoring Safeguard Systems

    International Nuclear Information System (INIS)

    Castleberry, K.; Lenarduzzi, R.; Whitaker, M.

    1999-01-01

    One of the several activities the International Atomic Energy Agency (IAEA) inspectors perform in the verification process of Safeguard operations is the review and correlation of data from different sources. This process is often complex due to the different forms in which the data is presented. This paper describes some of the elements that are necessary to create a ''standardized'' structure for the verification of data. When properly collected and formatted, data can be analyzed with off-the shelf software applications using customized macros to automate the commands for the desired analysis. The standardized-data collection methodology is based on instrumentation guidelines as well as data structure elements, such as verifiable timing of data entry, automated data logging, identification codes, and others. The identification codes are used to associate data items with their sources and to correlate them with items from other data logging activities. The addition of predefined parameter ranges allows automated evaluation with the capability to provide a data summary, a cross-index of all data related to a specific event. Instances of actual databases are used as examples. The data collection guidelines described in this paper facilitate the use of data from a variety of instrumentation platforms and also allow the instrumentation itself to be more easily applied in subsequent monitoring applications

  13. Automatic heating control system

    Energy Technology Data Exchange (ETDEWEB)

    Whittle, A.J.

    1989-11-15

    A heating control system for buildings comprises at least one heater incorporating heat storage means, a first sensor for detecting temperature within the building, means for setting a demand temperature, a second sensor for detecting outside temperature, a timer, and means for determining the switch on time of the heat storage means on the basis of the demand temperature and the internal and external temperatures. The system may additionally base the switch on time of the storage heater(s) on the heating and cooling rates of the building (as determined from the sensed temperatures); or on the anticipated daytime temperature (determined from the sensed night time temperature). (author).

  14. Automatic system for ionization chamber current measurements

    International Nuclear Information System (INIS)

    Brancaccio, Franco; Dias, Mauro S.; Koskinas, Marina F.

    2004-01-01

    The present work describes an automatic system developed for current integration measurements at the Laboratorio de Metrologia Nuclear of Instituto de Pesquisas Energeticas e Nucleares. This system includes software (graphic user interface and control) and a module connected to a microcomputer, by means of a commercial data acquisition card. Measurements were performed in order to check the performance and for validating the proposed design

  15. A microprogrammable high-speed data collection system for position sensitive X-ray detectors

    International Nuclear Information System (INIS)

    Hashizume, H.

    1984-01-01

    A high-speed data acquisition system has been designed which collects digital data from one- and two-dimensional position sensitive X-ray detectors at a maximum average data rate of 1 MHz. The system consists of two separate fast buffer memories, a 64 K word by 20-bit main storage, two timers, a display controller, a computer interface and a keyboard, controlled by a specially designed microprogrammable microprocessor. Data collection is performed by executing a microprogram stored in the control storage; data coming from a detector are first accumulated in a small but fast buffer memory by hardware and transferred to the main storage under control of the microprogram. This design not only permits time-resolved data collections but also provides maximum speed, flexibility and cost-effectiveness simultaneously. The system also accepts data from integrated detectors such as TV cameras. The system has been designed for use in experiments at conventional and synchrotron X-ray sources. (orig.)

  16. Automatic rebalancing of data in ATLAS distributed data management

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00473067; The ATLAS collaboration; Serfon, Cedric; Garonne, Vincent; Lassnig, Mario; Beermann, Thomas; Javurek, Tomas

    2017-01-01

    The ATLAS Distributed Data Management system stores more than 220PB of physics data across more than 130 sites globally. Rucio, the next generation data management system of the ATLAS collaboration, has now been successfully operated for two years. However, with the increasing workload and utilization, more automated and advanced methods of managing the data are needed. In this article we present an extension to the data management system, which is in charge of detecting and foreseeing storage elements reaching and surpassing their capacity limit. The system automatically and dynamically rebalances the data to other storage elements, while respecting and guaranteeing data distribution policies and ensuring the availability of the data. This concept not only lowers the operational burden, as these cumbersome procedures had previously to be done manually, but it also enables the system to use its distributed resources more efficiently, which not only affects the data management system itself, but in consequence...

  17. Automatic rebalancing of data in ATLAS distributed data management

    CERN Document Server

    Barisits, Martin-Stefan; The ATLAS collaboration; Garonne, Vincent; Lassnig, Mario; Beermann, Thomas

    2016-01-01

    The ATLAS Distributed Data Management system stores more than 220PB of physics data across more than 130 sites globally. Rucio, the next generation data management system of the ATLAS collaboration has now been successfully operated for over a year. However, with the forthcoming start of run-2 and its expected workload and utilization, more automated and advanced methods of managing the data are needed. In this article we present an extension to the data management system, which is in charge of detecting and foreseeing data imbalances as well as storage elements reaching and surpassing their capacity limit. The system automatically and dynamically rebalances the data to other storage elements, while respecting and guaranteeing data distribution policies and ensuring the availability of the data. This concept not only lowers the operational burden, as these cumbersome procedures had previously to be done manually, but it also enables the system to use its distributed resources more efficiently, which not only ...

  18. Development of automatic reactor vessel inspection systems: development of data acquisition and analysis system for the nuclear vessel weld

    Energy Technology Data Exchange (ETDEWEB)

    Park, C. H.; Lim, H. T.; Um, B. G. [Korea Advanced Institute of Science and Technology, Taejeon (Korea)

    2001-03-01

    The objective of this project is to develop an automated ultrasonic data acquisition and data analysis system to examine the reactor vessel weldsIn order to examine nuclear vessel welds including reactor pressure vessel(RPV), huge amount of ultrasonic data from 6 channels should be able to be on-line processed. In addition, ultrasonic transducer scanning device should be remotely controlled, because working place is high radiation area. This kind of an automated ultrasonic testing equipment has not been developed domestically yet In order to develop an automated ultrasonic testing system, RPV ultrasonic testing equipments developed in foreign countries were investigated and the capability of high speed ultrasonic signal processing hardwares was analyzed in this study, ultrasonic signal processing system was designed. And also, ultrasonic data acquisition and analysis software was developed. 11 refs., 6 figs., 9 tabs. (Author)

  19. The LANSCE (Los Alamos Neutron Scattering Center) target data collection system

    International Nuclear Information System (INIS)

    Kernodle, A.K.

    1989-01-01

    The Los Alamos Neutron Scattering Center (LANSCE) Target Data Collection System is the result of an effort to provide a base of information from which to draw conclusions on the performance and operational condition of the overall LANSCE target system. During the conceptualization of the system, several goals were defined. A survey was made of both custom-made and off-the-shelf hardware and software that were capable of meeting these goals. The first stage of the system was successfully implemented for the LANSCE run cycle 52. From the operational experience gained thus far, it appears that the LANSCE Target Data Collection System will meet all of the previously defined requirements

  20. A CityGML extension for traffic-sign objects that guides the automatic processing of data collected using Mobile Mapping technology

    Science.gov (United States)

    Varela-González, M.; Riveiro, B.; Arias-Sánchez, P.; González-Jorge, H.; Martínez-Sánchez, J.

    2014-11-01

    The rapid evolution of integral schemes, accounting for geometric and semantic data, has been importantly motivated by the advances in the last decade in mobile laser scanning technology; automation in data processing has also recently influenced the expansion of the new model concepts. This paper reviews some important issues involved in the new paradigms of city 3D modelling: an interoperable schema for city 3D modelling (cityGML) and mobile mapping technology to provide the features that composing the city model. This paper focuses in traffic signs, discussing their characterization using cityGML in order to ease the implementation of LiDAR technology in road management software, as well as analysing some limitations of the current technology in the labour of automatic detection and classification.

  1. Calibration of automatic performance measures - speed and volume data: volume 2, evaluation of the accuracy of approach volume counts and speeds collected by microwave sensors.

    Science.gov (United States)

    2016-05-01

    This study evaluated the accuracy of approach volumes and free flow approach speeds collected by the Wavetronix : SmartSensor Advance sensor for the Signal Performance Metrics system of the Utah Department of Transportation (UDOT), : using the field ...

  2. Portable data collection terminal in the automated power consumption measurement system

    Science.gov (United States)

    Vologdin, S. V.; Shushkov, I. D.; Bysygin, E. K.

    2018-01-01

    Aim of efficiency increasing, automation process of electric energy data collection and processing is very important at present time. High cost of classic electric energy billing systems prevent from its mass application. Udmurtenergo Branch of IDGC of Center and Volga Region developed electronic automated system called “Mobile Energy Billing” based on data collection terminals. System joins electronic components based on service-oriented architecture, WCF services. At present time all parts of Udmurtenergo Branch electric network are connected to “Mobile Energy Billing” project. System capabilities are expanded due to flexible architecture.

  3. Development of automatic reactor vessel inspection systems; development of data acquisition and analysis system for the nuclear vessel weld

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Jong Po; Park, C. H.; Kim, H. T.; Noh, H. C.; Lee, J. M.; Kim, C. K.; Um, B. G. [Research Institute of KAITEC, Seoul (Korea)

    2002-03-01

    The objective of this project is to develop an automated ultrasonic data acquisition and data analysis system to examine heavy vessel welds. In order to examine nuclear vessel welds including reactor pressure vessel(RPV), huge amount of ultrasonic data from 6 channels should be able to be on-line processed. In addition, ultrasonic transducer scanning device should be remotely controlled, because working place is high radiation area. This kind of an automated ultrasonic testing equipment has not been developed domestically yet. In order to develop an automated ultrasonic testing system, RPV ultrasonic testing equipments developed in foreign countries were investigated and the capability of high speed ultrasonic signal processing hardwares was analyzed. In this study, ultrasonic signal processing system was designed. And also, ultrasonic data acquisition software was developed. The new systems were tested on the RPV welds of Ulchin Unit 6 to confirm their functions and capabilities. They worked very well as designed and the tests were successfully completed. 13 refs., 34 figs., 11 tabs. (Author)

  4. Progress on statistical learning systems as data mining tools for the creation of automatic databases in Fusion environments

    International Nuclear Information System (INIS)

    Vega, J.; Murari, A.; Ratta, G.A.; Gonzalez, S.; Dormido-Canto, S.

    2010-01-01

    Nowadays, processing all information of a fusion database is a much more important issue than acquiring more data. Although typically fusion devices produce tens of thousands of discharges, specialized databases for physics studies are normally limited to a few tens of shots. This is due to the fact that these databases are almost always generated manually, which is a very time consuming and unreliable activity. The development of automatic methods to create specialized databases ensures first, the reduction of human efforts to identify and locate physical events, second, the standardization of criteria (reducing the vulnerability to human errors) and, third, the improvement of statistical relevance. Classification and regression techniques have been used for these purposes. The objective has been the automatic recognition of physical events (that can appear in a random and/or infrequent way) in waveforms and video-movies. Results are shown for the JET database.

  5. Ground truth data collection on mining industrial explosions registered by the International Monitoring System

    International Nuclear Information System (INIS)

    Ehl'tekov, A.Yu.; Gordon, V.P.; Firsov, V.A.; Chervyakov, V.B.

    2004-01-01

    The presentation is dedicated to organizational and technical issues connected with the task of Comprehensive Test-Ban-Treaty Organization timely notification on large chemical explosions including data on explosion location and time, on applied explosive substance quantity and type, and also on configuration and assumed purpose of explosion. Explosions registered by International Monitoring System are of special interest. Their data could be used for calibration of the monitoring system. Ground truth data collection and some explosions location results on Russia's mining enterprises were given. Ground truth data collection peculiarities according to mining industrial explosions were considered. (author)

  6. Data mining to detect clinical mastitis with automatic milking

    NARCIS (Netherlands)

    Kamphuis, C.; Mollenhorst, H.; Heesterbeek, J.A.P.; Hogeveen, H.

    2010-01-01

    Our objective was to use data mining to develop and validate a detection model for clinical mastitis (CM) using sensor data collected at nine Dutch dairy herds milking automatically. Sensor data was available for almost 3.5 million quarter milkings (QM) from 1,109 cows; 348 QM with CM were observed

  7. [Automatic adjustment control system for DC glow discharge plasma source].

    Science.gov (United States)

    Wan, Zhen-zhen; Wang, Yong-qing; Li, Xiao-jia; Wang, Hai-zhou; Shi, Ning

    2011-03-01

    There are three important parameters in the DC glow discharge process, the discharge current, discharge voltage and argon pressure in discharge source. These parameters influence each other during glow discharge process. This paper presents an automatic control system for DC glow discharge plasma source. This system collects and controls discharge voltage automatically by adjusting discharge source pressure while the discharge current is constant in the glow discharge process. The design concept, circuit principle and control program of this automatic control system are described. The accuracy is improved by this automatic control system with the method of reducing the complex operations and manual control errors. This system enhances the control accuracy of glow discharge voltage, and reduces the time to reach discharge voltage stability. The glow discharge voltage stability test results with automatic control system are provided as well, the accuracy with automatic control system is better than 1% FS which is improved from 4% FS by manual control. Time to reach discharge voltage stability has been shortened to within 30 s by automatic control from more than 90 s by manual control. Standard samples like middle-low alloy steel and tin bronze have been tested by this automatic control system. The concentration analysis precision has been significantly improved. The RSDs of all the test result are better than 3.5%. In middle-low alloy steel standard sample, the RSD range of concentration test result of Ti, Co and Mn elements is reduced from 3.0%-4.3% by manual control to 1.7%-2.4% by automatic control, and that for S and Mo is also reduced from 5.2%-5.9% to 3.3%-3.5%. In tin bronze standard sample, the RSD range of Sn, Zn and Al elements is reduced from 2.6%-4.4% to 1.0%-2.4%, and that for Si, Ni and Fe is reduced from 6.6%-13.9% to 2.6%-3.5%. The test data is also shown in this paper.

  8. Commutated automatic gain control system

    Science.gov (United States)

    Yost, S. R.

    1982-01-01

    A commutated automatic gain control (AGC) system was designed and built for a prototype Loran C receiver. The receiver uses a microcomputer to control a memory aided phase-locked loop (MAPLL). The microcomputer also controls the input/output, latitude/longitude conversion, and the recently added AGC system. The circuit designed for the AGC is described, and bench and flight test results are presented. The AGC circuit described actually samples starting at a point 40 microseconds after a zero crossing determined by the software lock pulse ultimately generated by a 30 microsecond delay and add network in the receiver front end envelope detector.

  9. Master data extraction and adaptation based on collected production data in manufacturing execution systems

    OpenAIRE

    Dimitrov, T.; Baumann, M.; Schenk, M.

    2010-01-01

    This paper presents an approach to extraction and correction of manufacturing master data, needed by Manufacturing Execution Systems (MES) to control and schedule the production. The implementation of the created schedule and the improvement of Key Performance Indicators depends strongly on the quality of the master data. The master data of most enterprises ages or the enterprises cannot fully provide it, because a highly manual expense for statistical analysis and administration is needed. T...

  10. A high-speed data-collection system for large-unit-cell crystals using an imaging plate as a detector

    International Nuclear Information System (INIS)

    Sato, Mamoru; Yamamoto, Masaki; Imada, Katsumi; Katsube, Yukiteru; Tanaka, Nobuo; Higashi, Tsuneyuki

    1992-01-01

    A high-speed data-collection system for large-unit-cell crystals is presented, using the Fuji Imaging Plate as an X-ray detector and a rotating-anode generator as the X-ray source. It is an automatic data-acquisition system that requires almost no manual intervention. The quality of data collected on the system is discussed. Merging R values ranged from 0.04 to 0.05. Compared with a four-circle diffractometer, data reproducibility was better, isomorphous/anomalous Patterson maps were almost identical in quality and data from a small-molecule crystal, cytidine, were of almost the same quality. Protein structures were refinable using the data measured on the system, the final crystallographic R value of the 2.2 A 3-isopropylmalate dehydrogenase structure being 0.185 and that of the 1.88 A Flammulina veltipes agglutinin structure being 0.199. (orig.)

  11. Development of a research-oriented system for collecting mechanical ventilator waveform data.

    Science.gov (United States)

    Rehm, Gregory B; Kuhn, Brooks T; Delplanque, Jean-Pierre; Guo, Edward C; Lieng, Monica K; Nguyen, Jimmy; Anderson, Nicholas R; Adams, Jason Y

    2017-10-28

    Lack of access to high-frequency, high-volume patient-derived data, such as mechanical ventilator waveform data, has limited the secondary use of these data for research, quality improvement, and decision support. Existing methods for collecting these data are obtrusive, require high levels of technical expertise, and are often cost-prohibitive, limiting their use and scalability for research applications. We describe here the development of an unobtrusive, open-source, scalable, and user-friendly architecture for collecting, transmitting, and storing mechanical ventilator waveform data that is generalizable to other patient care devices. The system implements a software framework that automates and enforces end-to-end data collection and transmission. A web-based data management application facilitates nontechnical end users' abilities to manage data acquisition devices, mitigates data loss and misattribution, and automates data storage. Using this integrated system, we have been able to collect ventilator waveform data from >450 patients as part of an ongoing clinical study. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  12. Antares automatic beam alignment system

    International Nuclear Information System (INIS)

    Appert, Q.; Swann, T.; Sweatt, W.; Saxman, A.

    1980-01-01

    Antares is a 24-beam-line CO 2 laser system for controlled fusion research, under construction at Los Alamos Scientific Laboratory (LASL). Rapid automatic alignment of this system is required prior to each experiment shot. The alignment requirements, operational constraints, and a developed prototype system are discussed. A visible-wavelength alignment technique is employed that uses a telescope/TV system to view point light sources appropriately located down the beamline. Auto-alignment is accomplished by means of a video centroid tracker, which determines the off-axis error of the point sources. The error is nulled by computer-driven, movable mirrors in a closed-loop system. The light sources are fiber-optic terminations located at key points in the optics path, primarily at the center of large copper mirrors, and remotely illuminated to reduce heating effects

  13. Technology for collecting and analyzing relational data

    Directory of Open Access Journals (Sweden)

    E. N. Fedorova

    2016-01-01

    Full Text Available Purpose of the study. The scientific and educational organizations use traditionally e-mail with Microsoft Excel spreadsheets and Microsoft Word documents for operational data collection. The disadvantages of this approach include the lack of control of the correctness of the data input, the complexity of processing the information received due to non-relational data model, etc. There are online services that enable to organize the collection of data in a relational form. The disadvantages of these systems are: the absence of thesaurus support; a limited set of elements of data input control; the limited control the operation of the input form; most of the systems is shareware, etc. Thus, it is required the development of Internet data collection and analysis technology, which should allow to identify quickly model the data collected and automatically implement data collection in accordance with this model.Materials and methods. The article describes the technology developed and tested for operational data collection and analysis using "Faramant" system. System operation "Faramant" is based on a model document, which includes three components: description of the data structure; visualization; logic of form work. All stages of the technology are performed by the user using the browser. The main stage of the proposed technology is the definition of the data model as a set of relational tables. To create a table within the system it’s required to determine the name and a list of fields. For each field, you must specify its name and use the control to the data input and logic of his work. Controls are used to organize the correct input data depending on the data type. Based on a model system "Faramant" automatically creates a filling form, using which users can enter information. To change the form visualization, you can use the form template. The data can be viewed page by page in a table. For table rows, you can apply different filters. To

  14. A Retrospective Review of the Clinical Characteristics and Blood Glucose Data from Cellnovo System Users using Data Collected from the Cellnovo Online Platform.

    Science.gov (United States)

    Hautier-Suply, Olivia; Friedmann, Yasmin; Shapley, Julian

    2018-04-01

    Technological advances have led to innovative insulin delivery systems for patients with type 1 diabetes mellitus. In particular, the combination of miniature engineering and software algorithms contained in continuous subcutaneous insulin infusion (CSII) system pumps provide the user and the healthcare practitioner with an opportunity to review and adjust blood glucose (BG) levels according to system feedback, and to modify or programme their regimen according to their needs. While CSII pumps record a number of data parameters such as BG level, carbohydrate intake, activity and insulin delivered, these data are generally 'locked in' and can only be accessed by uploading to a cloud-based system, thus information is not contemporaneous. The Cellnovo Diabetes Management System (Cellnovo, Bridgend, UK) allows data to be transmitted securely and wirelessly in real time to a secure server, which is then retrieved by an online platform, the Cellnovo Online platform, enabling continuous access by the user and by clinicians. In this article, the authors describe a retrospective review of the patient data automatically uploaded to the Cellnovo Online platform. Baseline clinical and demographic characteristics collected at the start of pump therapy are shown for all patients, and BG data from a sub-cohort of patients who have been using the system for at least 6 months and who take and record an average of three BG level tests per day are presented to demonstrate glycaemic data over time.

  15. Development of user-friendly and interactive data collection system for cerebral palsy.

    Science.gov (United States)

    Raharjo, I; Burns, T G; Venugopalan, J; Wang, M D

    2016-02-01

    Cerebral palsy (CP) is a permanent motor disorder that appears in early age and it requires multiple tests to assess the physical and mental capabilities of the patients. Current medical record data collection systems, e.g., EPIC, employed for CP are very general, difficult to navigate, and prone to errors. The data cannot easily be extracted which limits data analysis on this rich source of information. To overcome these limitations, we designed and prototyped a database with a graphical user interface geared towards clinical research specifically in CP. The platform with MySQL and Java framework is reliable, secure, and can be easily integrated with other programming languages for data analysis such as MATLAB. This database with GUI design is a promising tool for data collection and can be applied in many different fields aside from CP to infer useful information out of the vast amount of data being collected.

  16. JIGSAW: Acquisition, Display and Analysis system designed to collect data from Multiple Gamma-Ray detectors

    International Nuclear Information System (INIS)

    Haywood, S.E.; Bamford, G.J.; Rester, A.C.; Coldwell, R.L.

    1992-01-01

    In this paper, the authors report on work performed to date on JIGSAW - a self contained data acquisition, display and analysis system designed to collect data form multiple gamma-ray detectors. The data acquisition system utilizes commercially available VMEbus and NIM hardware modules and the VME exec real time operating system. A Unix based software package, written in ANSI standard C and with the XII graphics routines, allows the user to view the acquired spectra. Analysis of the histograms can be performed in background during the run with the ROBFIT suite of curve fitting routines

  17. CHLOE: a system for the automatic handling of spark pictures

    International Nuclear Information System (INIS)

    Butler, J.W.; Hodges, D.; Royston, R.

    The system for automatic data handling uses commercially available or state-of-the-art components. The system is flexible enough to accept information from various types of experiments involving photographic data acquisition

  18. Automatic TLI recognition system. Part 1: System description

    Energy Technology Data Exchange (ETDEWEB)

    Partin, J.K.; Lassahn, G.D.; Davidson, J.R.

    1994-05-01

    This report describes an automatic target recognition system for fast screening of large amounts of multi-sensor image data, based on low-cost parallel processors. This system uses image data fusion and gives uncertainty estimates. It is relatively low cost, compact, and transportable. The software is easily enhanced to expand the system`s capabilities, and the hardware is easily expandable to increase the system`s speed. This volume gives a general description of the ATR system.

  19. Automatization of hydrodynamic modelling in a Floreon+ system

    Science.gov (United States)

    Ronovsky, Ales; Kuchar, Stepan; Podhoranyi, Michal; Vojtek, David

    2017-07-01

    The paper describes fully automatized hydrodynamic modelling as a part of the Floreon+ system. The main purpose of hydrodynamic modelling in the disaster management is to provide an accurate overview of the hydrological situation in a given river catchment. Automatization of the process as a web service could provide us with immediate data based on extreme weather conditions, such as heavy rainfall, without the intervention of an expert. Such a service can be used by non scientific users such as fire-fighter operators or representatives of a military service organizing evacuation during floods or river dam breaks. The paper describes the whole process beginning with a definition of a schematization necessary for hydrodynamic model, gathering of necessary data and its processing for a simulation, the model itself and post processing of a result and visualization on a web service. The process is demonstrated on a real data collected during floods in our Moravian-Silesian region in 2010.

  20. Automatic continuous monitoring system for dangerous sites and cargoes

    International Nuclear Information System (INIS)

    Smirnov, S.N.

    2009-01-01

    The problems of creation of automatic comprehensive continuous monitoring system for nuclear and radiation sites and cargoes of Rosatom Corporation, which carries out data collecting, processing, storage and transmission, including informational support to decision-making, as well as support to modelling and forecasting functions, are considered. The system includes components of two levels: site and industry. Currently the system is used to monitor over 8000 integrated parameters, which characterise the status of nuclear and radiation safety on Rosatom sites, environmental and fire safety

  1. Real-time environmental radiation monitoring system with automatic restoration of backup data in site detector via communication using radio frequency

    International Nuclear Information System (INIS)

    Lee, Wan No; Kim, Eun Han; Chung, Kun Ho; Cho, Young Hyun; Choi, Geun Sik; Lee, Chang Woo; Park, Ki Hyun; Kim, Yun Goo

    2003-01-01

    An environmental radiation monitoring system based on high pressurized ionization chamber has been used for on-line gamma monitoring surrounding the KAERI (Korea Atomic Energy Research Institute), which transmits the dose data measured from ion chamber on the site via radio frequency to a central processing computer and stores the transmitted real-time data. Although communication using ratio frequency has several advantages such as effective and economical transmission, storage, and data process, there is one main disadvantage that data loss during transmission often happens because of unexpected communication problems. It is possible to restore the loss data by off-line such as floppy disk but the simultaneous process and display of current data as well as the backup data are very difficult in the present on-line system. In this work, a new electronic circuit board and the operation software applicable to the conventional environmental radiation monitoring system are developed and the automatical synchronization of the ion chamber unit and the central processing computer is carried out every day. This system is automatically able to restore the backup data within 34 hours without additional equipment and also display together the current data as well as the transmitted backup data after checking time flag

  2. A Machine Vision System for Automatically Grading Hardwood Lumber - (Proceedings)

    Science.gov (United States)

    Richard W. Conners; Tai-Hoon Cho; Chong T. Ng; Thomas H. Drayer; Joe G. Tront; Philip A. Araman; Robert L. Brisbon

    1990-01-01

    Any automatic system for grading hardwood lumber can conceptually be divided into two components. One of these is a machine vision system for locating and identifying grading defects. The other is an automatic grading program that accepts as input the output of the machine vision system and, based on these data, determines the grade of a board. The progress that has...

  3. Automatic mapping of monitoring data

    DEFF Research Database (Denmark)

    Lophaven, Søren; Nielsen, Hans Bruun; Søndergaard, Jacob

    2005-01-01

    of identifying a release of radioactivity provided that the number of sampling locations is sufficiently high. Consequently, we believe that a combination of applying the presented mapping approach and the physical knowledge of the transport processes of radioactivity should be used to predict the extreme values.......). In the second dataset an accidental release of radioactivity in the environment was simulated in the South-Western corner of the monitored area. The approach has a tendency to smooth the actual data values, and therefore it underestimates extreme values, as seen in the second dataset. However, it is capable...

  4. A mobile field-work data collection system for the wireless era of health surveillance.

    Science.gov (United States)

    Forsell, Marianne; Sjögren, Petteri; Renard, Matthew; Johansson, Olle

    2011-03-01

    In many countries or regions the capacity of health care resources is below the needs of the population and new approaches for health surveillance are needed. Innovative projects, utilizing wireless communication technology, contribute to reliable methods for field-work data collection and reporting to databases. The objective was to describe a new version of a wireless IT-support system for field-work data collection and administration. The system requirements were drawn from the design objective and translated to system functions. The system architecture was based on fieldwork experiences and administrative requirements. The Smartphone devices were HTC Touch Diamond2s, while the system was based on a platform with Microsoft .NET components, and a SQL Server 2005 with Microsoft Windows Server 2003 operating system. The user interfaces were based on .NET programming, and Microsoft Windows Mobile operating system. A synchronization module enabled download of field data to the database, via a General Packet Radio Services (GPRS) to a Local Area Network (LAN) interface. The field-workers considered the here-described applications user-friendly and almost self-instructing. The office administrators considered that the back-office interface facilitated retrieval of health reports and invoice distribution. The current IT-support system facilitates short lead times from fieldwork data registration to analysis, and is suitable for various applications. The advantages of wireless technology, and paper-free data administration need to be increasingly emphasized in development programs, in order to facilitate reliable and transparent use of limited resources.

  5. Use of the BOINC system for distributed data collection in the “Internet of Things”

    Directory of Open Access Journals (Sweden)

    Handrik Marián

    2018-01-01

    Full Text Available The BOINC software package allows the creation of massive computational systems for distributed computing. The system is designed to use heterogeneous computing devices that are differing in hardware architecture and operating systems. The variability and versatility of the BOINC system allows its use to create a distributed system for measurement and data collection. This is done through the standard HTTP and HTTPS web communication protocol that is used on the Internet in connected devices, so called “Internet of things”.

  6. A technical assistance on data collection on subdivision of wet-system apparatuses

    International Nuclear Information System (INIS)

    2001-09-01

    In the Ningyo-Toge Environmental Engineering Center, development on subdivision engineering system for abolition of nuclear fuel facilities has been promoted. However, data on subdivision of instruments and apparatuses to be carried out as a part of the abolition was insufficient. Therefore, here was intended to investigate data collections so as to use subdivision of a wet-system apparatuses of the smelting conversion facility begun on June, 2000, as a field of data collection as effectively as possible, on construction of the system rationally supporting abolition of nuclear fuel facility promoted at the Ningyo-Toge Environmental Engineering Center. This subdivision of the wet-system apparatuses of the facility is programmed to carry out the subdivision for two years of 2000 and 2001 fiscal years. Its working procedure is begun from non-polluted matters (electrics, instruments, and utility pipings) at every rooms to carry out appliances using uranium. Here were reported on present states survey of the subdivision, kinds and frequencies of data at the subdivision, data collection manual, and rationalization of data recording method. (G.K.)

  7. Imitation simulation of collective acquisition systems for nuclear-physical data

    International Nuclear Information System (INIS)

    Ryabov, Yu.F.; Khomutnikov, V.P.

    1981-01-01

    The imitation simulation method of collective investigation systems (CIS) oriented to the acquisition and express-processing of experiments performed on the basis of multi-beam research reactors under film data measurements by means of semiautomatic installations is described. The method is based on description of specific requests to the system as service phase sequences (small computer data acquisition, mass transfer in CIS, data bank record processing). The algorithm basis of these phases performed is delay imitation at each of the means employed. The structure of the simulating program is presented. The complex of programs is realized on the base of the NOST program packet for the M-4030 computer. Its efficiency in collective multiprogram systems investigation is shown [ru

  8. The nuclear data collecting system designed with ARM and μC/OS-II

    International Nuclear Information System (INIS)

    Wang Chunsheng; Ma Yingjie; Han Feng

    2008-01-01

    Introduce a kind of nuclear data collecting system regarding ARM-μC/OS-II as the platform, gathering the GPS receiver in it. It was detailed to expatiated the key techniques of the multi-channel pulse amplitude analyzer, the interface design of LPC2148, a controller in ARM, and how to expand the RTOS and design applications on μC/OS-II. This system can communicate with the GPS-OEM module by the UART interface, collecting the GPS information synchronously as well as nuclear data. And then save and display them or pass them to the host computer by the USB interface. The embedded and Real-Time system, μC/OS-II build up the real-time and stability of the system and advance the integration. (authors)

  9. Integrated straight - through automatic non-destructive examination and data acquisition system for thin-wall tubes

    International Nuclear Information System (INIS)

    Stoessel, A.; Boulanger, G.; Furlan, J.; Mogavero, R.

    1981-09-01

    This non-destructive testing unit inspects the cladding tubes for the SUPER-PHENIX fast neutron reactor. The quality level demanded for these tubes, as well as their number, required designing an installation that combined high performance with a great testing rate and complete automation. The testing is effected under immersion by means of six transducers, focused in line, working at 30 MHz. The tubes are numbered on an automatic rig; marking is by dark rings obtained by superficial electrolysis of the tube and regularly distributed on the abscissa; the quality of the tube is not affected by this. The advantage of this numbering system is that it enables the tubes to be fed to the test set in any order. An acquisition unit, constituted of a microprocessor, a semi-graphical printer and a double floppy disk unit, makes it possible to enter, edit and store the information for each tube [fr

  10. Development of Inspection Data Collection and Evaluation System (IDES) for J-MOX (1)

    International Nuclear Information System (INIS)

    Kumakura, Shinichi; Takizawa, Koji; Masuda, Shoichiro; Iso, Shoko; Kikuchi, Masahiro; Hisamatsu, Yoshinori; Kurobe, Hiroko; Kawasue, Akane

    2012-01-01

    'Inspection Data and Collection and Evaluation System' is the system to storage inspection data and operator declaration data collected from various measurement equipments, which are installed in fuel fabrication processes of the large-scale MOX fuel fabrication plant, and to make safeguards evaluation using these data. Nuclear Material Control Center is now developing this system under the project commissioned by JSGO. By last fiscal year, we developed the simulator to simulate fuel fabrication process and generate data simulating in-process material inventory/flow and these measurement data. In addition, we developed a verification evaluation system to calculate various statistics from the simulation data and conduct statistical tests such as NRTA in order to verify the adequacy of material accountancy for the fabrication process. We are currently investigating the adequacy of evaluation itself and effects for evaluation by changing various process factors including unmeasured inventories as well as the adequacy of current safeguards approach. In the presentation, we explain the developed system configuration, calculation method of the simulation etc. and demonstrate same examples of the simulated result on material flow in the fabrication process and a part of the analytical results. (author)

  11. Development of an automatic reactor inspection system

    International Nuclear Information System (INIS)

    Kim, Jae Hee; Eom, Heung Seop; Lee, Jae Cheol; Choi, Yoo Raek; Moon, Soon Seung

    2002-02-01

    Using recent technologies on a mobile robot computer science, we developed an automatic inspection system for weld lines of the reactor vessel. The ultrasonic inspection of the reactor pressure vessel is currently performed by commercialized robot manipulators. Since, however, the conventional fixed type robot manipulator is very huge, heavy and expensive, it needs long inspection time and is hard to handle and maintain. In order to resolve these problems, we developed a new automatic inspection system using a small mobile robot crawling on the vertical wall of the reactor vessel. According to our conceptual design, we developed the reactor inspection system including an underwater inspection robot, a laser position control subsystem, an ultrasonic data acquisition/analysis subsystem and a main control subsystem. We successfully carried out underwater experiments on the reactor vessel mockup, and real reactor ready for Ulchine nuclear power plant unit 6 at Dusan Heavy Industry in Korea. After this project, we have a plan to commercialize our inspection system. Using this system, we can expect much reduction of the inspection time, performance enhancement, automatic management of inspection history, etc. In the economic point of view, we can also expect import substitution more than 4 million dollars. The established essential technologies for intelligent control and automation are expected to be synthetically applied to the automation of similar systems in nuclear power plants

  12. Research on the Method of Big Data Collecting, Storing and Analyzing of Tongue Diagnosis System

    Science.gov (United States)

    Chen, Xiaowei; Wu, Qingfeng

    2018-03-01

    This paper analyzes the contents of the clinical data of tongue diagnosis of TCM (Traditional Chinese Medicine), and puts forward a method to collect, store and analyze the clinical data of tongue diagnosis. Under the guidance of TCM theory of syndrome differentiation and treatment, this method combines with Hadoop, which is a distributed computing system with strong expansibility, and integrates the functions of analysis and conversion of big data of clinic tongue diagnosis. At the same time, the consistency, scalability and security of big data in tongue diagnosis are realized.

  13. Steps towards single source--collecting data about quality of life within clinical information systems.

    Science.gov (United States)

    Fritz, Fleur; Ständer, Sonja; Breil, Bernhard; Dugas, Martin

    2010-01-01

    Information about the quality of life from patients being treated in routine medical care is important for the attending physician. This data is also needed in research for example to evaluate the therapy and the course of the disease respectively. Especially skin diseases often negatively affect the quality of life. Therefore we aimed to design a concept to collect such data during treatment and use it for both medical care and research in the setting of dermatology. We performed a workflow analysis and implemented a designated form using the tools of the local clinical information system. Quality of life data is now collected within the clinical information system during treatment and is used for discharge letters, progress overviews as well as research about the treatment and course of disease. This concept which contributes to the single source approach was feasible within dermatology and is ready to be expanded into other domains.

  14. Semi-automatic Data Integration using Karma

    Science.gov (United States)

    Garijo, D.; Kejriwal, M.; Pierce, S. A.; Houser, P. I. Q.; Peckham, S. D.; Stanko, Z.; Hardesty Lewis, D.; Gil, Y.; Pennington, D. D.; Knoblock, C.

    2017-12-01

    Data integration applications are ubiquitous in scientific disciplines. A state-of-the-art data integration system accepts both a set of data sources and a target ontology as input, and semi-automatically maps the data sources in terms of concepts and relationships in the target ontology. Mappings can be both complex and highly domain-specific. Once such a semantic model, expressing the mapping using community-wide standard, is acquired, the source data can be stored in a single repository or database using the semantics of the target ontology. However, acquiring the mapping is a labor-prone process, and state-of-the-art artificial intelligence systems are unable to fully automate the process using heuristics and algorithms alone. Instead, a more realistic goal is to develop adaptive tools that minimize user feedback (e.g., by offering good mapping recommendations), while at the same time making it intuitive and easy for the user to both correct errors and to define complex mappings. We present Karma, a data integration system that has been developed over multiple years in the information integration group at the Information Sciences Institute, a research institute at the University of Southern California's Viterbi School of Engineering. Karma is a state-of-the-art data integration tool that supports an interactive graphical user interface, and has been featured in multiple domains over the last five years, including geospatial, biological, humanities and bibliographic applications. Karma allows a user to import their own ontology and datasets using widely used formats such as RDF, XML, CSV and JSON, can be set up either locally or on a server, supports a native backend database for prototyping queries, and can even be seamlessly integrated into external computational pipelines, including those ingesting data via streaming data sources, Web APIs and SQL databases. We illustrate a Karma workflow at a conceptual level, along with a live demo, and show use cases of

  15. Solar Powered Automatic Shrimp Feeding System

    Directory of Open Access Journals (Sweden)

    Dindo T. Ani

    2015-12-01

    Full Text Available - Automatic system has brought many revolutions in the existing technologies. One among the technologies, which has greater developments, is the solar powered automatic shrimp feeding system. For instance, the solar power which is a renewable energy can be an alternative solution to energy crisis and basically reducing man power by using it in an automatic manner. The researchers believe an automatic shrimp feeding system may help solve problems on manual feeding operations. The project study aimed to design and develop a solar powered automatic shrimp feeding system. It specifically sought to prepare the design specifications of the project, to determine the methods of fabrication and assembly, and to test the response time of the automatic shrimp feeding system. The researchers designed and developed an automatic system which utilizes a 10 hour timer to be set in intervals preferred by the user and will undergo a continuous process. The magnetic contactor acts as a switch connected to the 10 hour timer which controls the activation or termination of electrical loads and powered by means of a solar panel outputting electrical power, and a rechargeable battery in electrical communication with the solar panel for storing the power. By undergoing through series of testing, the components of the modified system were proven functional and were operating within the desired output. It was recommended that the timer to be used should be tested to avoid malfunction and achieve the fully automatic system and that the system may be improved to handle changes in scope of the project.

  16. Data collection and storage in long-term ecological and evolutionary studies: The Mongoose 2000 system.

    Science.gov (United States)

    Marshall, Harry H; Griffiths, David J; Mwanguhya, Francis; Businge, Robert; Griffiths, Amber G F; Kyabulima, Solomon; Mwesige, Kenneth; Sanderson, Jennifer L; Thompson, Faye J; Vitikainen, Emma I K; Cant, Michael A

    2018-01-01

    Studying ecological and evolutionary processes in the natural world often requires research projects to follow multiple individuals in the wild over many years. These projects have provided significant advances but may also be hampered by needing to accurately and efficiently collect and store multiple streams of the data from multiple individuals concurrently. The increase in the availability and sophistication of portable computers (smartphones and tablets) and the applications that run on them has the potential to address many of these data collection and storage issues. In this paper we describe the challenges faced by one such long-term, individual-based research project: the Banded Mongoose Research Project in Uganda. We describe a system we have developed called Mongoose 2000 that utilises the potential of apps and portable computers to meet these challenges. We discuss the benefits and limitations of employing such a system in a long-term research project. The app and source code for the Mongoose 2000 system are freely available and we detail how it might be used to aid data collection and storage in other long-term individual-based projects.

  17. Development of a system for data collection and processing by telemetry

    International Nuclear Information System (INIS)

    Tavares Filho, R.F.

    1983-01-01

    The environmental impact of nuclear industry is, obviously, a matter of the greatest concern. On account of that, a large number of parameters must be recorded during long periods with a high level of confidence. The site selection of brazilian nuclear power plants is conducted under this philosophy. Data acquisition of ocean related parameters in remote, non explored, areas is rather stringent. In order to avoid a series of problems with data collection and processing, a telemetric system concept was developed. Electronic aspects of this system are, mainly, emphasized. For such purpose the system is splitted into two sub-systems: the former for data collection, signal conditionning and transmission and the latter for signal reception and treatment. All parts of the systems were tested in the laboratory before an integrated check, the corresponding results being encouraging. The whole equipment was installed one year ago at the sea shore region of Peruibe, state of Sao Paulo, and is in operation, adequately, eversince. (Author) [pt

  18. A continuous flow from sample collection to data acceptability determination using an automated system

    International Nuclear Information System (INIS)

    Fisk, J.F.; Leasure, C.; Sauter, A.D.

    1993-01-01

    In its role as regulator, EPA is the recipient of enormous reams of analytical data, especially within the Superfund Program. In order to better manage the volume of paper that comes in daily, Superfund has required its laboratories to provide data that is contained on reporting forms to be delivered also on a diskette for uploading into data bases for various purposes, such as checking for contractual compliance, tracking quality assurance parameters, and, ultimately, for reviewing the data by computer. This last area, automated review of the data, has generated programs that are not necessarily appropriate for use by clients other than Superfund. Such is the case with Los Alamos National Laboratory's Environmental Chemistry Group and its emerging subcontractor community, designed to meet the needs of the remedial action program at LANL. LANL is in the process of implementing an automated system that will be used from the planning stage of sample collection to the production of a project-specific report on analytical data quality. Included are electronic scheduling and tracking of samples, data entry, checking and transmission, data assessment and qualification for use, and report generation that will tie the analytical data quality back to the performance criteria defined prior to sample collection. Industry standard products will be used (e.g., ORACLE, Microsoft Excel) to ensure support for users, prevent dependence on proprietary software, and to protect LANL's investment for the future

  19. 76 FR 58301 - Proposed Extension of Existing Information Collection; Automatic Fire Sensor and Warning Device...

    Science.gov (United States)

    2011-09-20

    ... Information Collection; Automatic Fire Sensor and Warning Device Systems; Examination and Test Requirements ACTION: Notice of request for public comments. SUMMARY: The Mine Safety and Health Administration (MSHA... public comment version of this information collection package. FOR FURTHER INFORMATION CONTACT: Roslyn B...

  20. [Development of automatic urine monitoring system].

    Science.gov (United States)

    Wei, Liang; Li, Yongqin; Chen, Bihua

    2014-03-01

    An automatic urine monitoring system is presented to replace manual operation. The system is composed of the flow sensor, MSP430f149 single chip microcomputer, human-computer interaction module, LCD module, clock module and memory module. The signal of urine volume is captured when the urine flows through the flow sensor and then displayed on the LCD after data processing. The experiment results suggest that the design of the monitor provides a high stability, accurate measurement and good real-time, and meets the demand of the clinical application.

  1. A cloud-based system for automatic glaucoma screening.

    Science.gov (United States)

    Fengshou Yin; Damon Wing Kee Wong; Ying Quan; Ai Ping Yow; Ngan Meng Tan; Gopalakrishnan, Kavitha; Beng Hai Lee; Yanwu Xu; Zhuo Zhang; Jun Cheng; Jiang Liu

    2015-08-01

    In recent years, there has been increasing interest in the use of automatic computer-based systems for the detection of eye diseases including glaucoma. However, these systems are usually standalone software with basic functions only, limiting their usage in a large scale. In this paper, we introduce an online cloud-based system for automatic glaucoma screening through the use of medical image-based pattern classification technologies. It is designed in a hybrid cloud pattern to offer both accessibility and enhanced security. Raw data including patient's medical condition and fundus image, and resultant medical reports are collected and distributed through the public cloud tier. In the private cloud tier, automatic analysis and assessment of colour retinal fundus images are performed. The ubiquitous anywhere access nature of the system through the cloud platform facilitates a more efficient and cost-effective means of glaucoma screening, allowing the disease to be detected earlier and enabling early intervention for more efficient intervention and disease management.

  2. Online Data Collection.

    Science.gov (United States)

    Topp, Neal W.; Pawloski, Bob

    2002-01-01

    Describes the eventful history of online data collection and presents a review of current literature following by a list of pros and cons to be considered when stepping into online surveying. (Contains 14 references.) (Author/YDS)

  3. Digital signal processing for CdTe detectors using VXIbus data collection systems

    Energy Technology Data Exchange (ETDEWEB)

    Fukuda, Daiji; Takahashi, Hiroyuki; Kurahashi, Tomohiko; Iguchi, Tetsuo; Nakazawa, Masaharu

    1996-07-01

    Recently fast signal digitizing technique has been developed, and signal waveforms with very short time periods can be obtained. In this paper, we analyzed each measured pulse which was digitized by an apparatus of this kind, and tried to improve an energy resolution of a CdTe semiconductor detector. The result of the energy resolution for {sup 137}Cs 662 keV photopeak was 13 keV. Also, we developed a fast data collection system based on VXIbus standard, and the counting rate on this system was obtained about 50 counts per second. (author)

  4. Improvement of an automated protein crystal exchange system PAM for high-throughput data collection

    International Nuclear Information System (INIS)

    Hiraki, Masahiko; Yamada, Yusuke; Chavas, Leonard M. G.; Wakatsuki, Soichi; Matsugaki, Naohiro

    2013-01-01

    A special liquid-nitrogen Dewar with double capacity for the sample-exchange robot has been created at AR-NE3A at the Photon Factory, allowing continuous fully automated data collection. In this work, this new system is described and the stability of its calibration is discussed. Photon Factory Automated Mounting system (PAM) protein crystal exchange systems are available at the following Photon Factory macromolecular beamlines: BL-1A, BL-5A, BL-17A, AR-NW12A and AR-NE3A. The beamline AR-NE3A has been constructed for high-throughput macromolecular crystallography and is dedicated to structure-based drug design. The PAM liquid-nitrogen Dewar can store a maximum of three SSRL cassettes. Therefore, users have to interrupt their experiments and replace the cassettes when using four or more of them during their beam time. As a result of investigation, four or more cassettes were used in AR-NE3A alone. For continuous automated data collection, the size of the liquid-nitrogen Dewar for the AR-NE3A PAM was increased, doubling the capacity. In order to check the calibration with the new Dewar and the cassette stand, calibration experiments were repeatedly performed. Compared with the current system, the parameters of the novel system are shown to be stable

  5. Improvement of an automated protein crystal exchange system PAM for high-throughput data collection

    Energy Technology Data Exchange (ETDEWEB)

    Hiraki, Masahiko, E-mail: masahiko.hiraki@kek.jp; Yamada, Yusuke; Chavas, Leonard M. G. [High Energy Accelerator Research Organization, 1-1 Oho, Tsukuba, Ibaraki 305-0801 (Japan); Wakatsuki, Soichi [High Energy Accelerator Research Organization, 1-1 Oho, Tsukuba, Ibaraki 305-0801 (Japan); SLAC National Accelerator Laboratory, 2575 Sand Hill Road, MS 69, Menlo Park, CA 94025-7015 (United States); Stanford University, Beckman Center B105, Stanford, CA 94305-5126 (United States); Matsugaki, Naohiro [High Energy Accelerator Research Organization, 1-1 Oho, Tsukuba, Ibaraki 305-0801 (Japan)

    2013-11-01

    A special liquid-nitrogen Dewar with double capacity for the sample-exchange robot has been created at AR-NE3A at the Photon Factory, allowing continuous fully automated data collection. In this work, this new system is described and the stability of its calibration is discussed. Photon Factory Automated Mounting system (PAM) protein crystal exchange systems are available at the following Photon Factory macromolecular beamlines: BL-1A, BL-5A, BL-17A, AR-NW12A and AR-NE3A. The beamline AR-NE3A has been constructed for high-throughput macromolecular crystallography and is dedicated to structure-based drug design. The PAM liquid-nitrogen Dewar can store a maximum of three SSRL cassettes. Therefore, users have to interrupt their experiments and replace the cassettes when using four or more of them during their beam time. As a result of investigation, four or more cassettes were used in AR-NE3A alone. For continuous automated data collection, the size of the liquid-nitrogen Dewar for the AR-NE3A PAM was increased, doubling the capacity. In order to check the calibration with the new Dewar and the cassette stand, calibration experiments were repeatedly performed. Compared with the current system, the parameters of the novel system are shown to be stable.

  6. An evaluation of semi-automated methods for collecting ecosystem-level data in temperate marine systems.

    Science.gov (United States)

    Griffin, Kingsley J; Hedge, Luke H; González-Rivero, Manuel; Hoegh-Guldberg, Ove I; Johnston, Emma L

    2017-07-01

    Historically, marine ecologists have lacked efficient tools that are capable of capturing detailed species distribution data over large areas. Emerging technologies such as high-resolution imaging and associated machine-learning image-scoring software are providing new tools to map species over large areas in the ocean. Here, we combine a novel diver propulsion vehicle (DPV) imaging system with free-to-use machine-learning software to semi-automatically generate dense and widespread abundance records of a habitat-forming algae over ~5,000 m 2 of temperate reef. We employ replicable spatial techniques to test the effectiveness of traditional diver-based sampling, and better understand the distribution and spatial arrangement of one key algal species. We found that the effectiveness of a traditional survey depended on the level of spatial structuring, and generally 10-20 transects (50 × 1 m) were required to obtain reliable results. This represents 2-20 times greater replication than have been collected in previous studies. Furthermore, we demonstrate the usefulness of fine-resolution distribution modeling for understanding patterns in canopy algae cover at multiple spatial scales, and discuss applications to other marine habitats. Our analyses demonstrate that semi-automated methods of data gathering and processing provide more accurate results than traditional methods for describing habitat structure at seascape scales, and therefore represent vastly improved techniques for understanding and managing marine seascapes.

  7. Geophysical data collection using an interactive personal computer system. Part 1. ; Experimental monitoring of Suwanosejima volcano

    Energy Technology Data Exchange (ETDEWEB)

    Iguchi, M. (Kyoto Univerdity, Kyoto (Japan). Disaster Prevention Reserach Institute)

    1991-10-15

    In the article, a computer-communication system was developed in order to collect geophysical data from remote volcanos via a public telephpne network. This system is composed of a host presonal computer at an observatory and several personal computers as terminals at remote stations. Each terminal acquires geophysical data, such as seismic, intrasonic, and ground deformation date. These gara are stored in the terminals temporarily, and transmitted to the host computer upon command from host computer. Experimental monitoring was conducted between Sakurajima Volcanological Observatory and several statins in the Satsunan Islands and southern Kyushu. The seismic and eruptive activities of Suwanosejima volcano were monitored by this system. Consequently, earthquakes and air-shocks accompanied by the explosive activity were observed. B-type earthquakes occurred prio to the relatively prolonged eruptive activity. Intermittent occurrences of volcanic tremors were also clearly recognized from the change in mean amplitubes of seismic waves. 7 refs., 10 figs., 2 tabs.

  8. ENT COBRA (Consortium for Brachytherapy Data Analysis: interdisciplinary standardized data collection system for head and neck patients treated with interventional radiotherapy (brachytherapy

    Directory of Open Access Journals (Sweden)

    Luca Tagliaferri

    2016-08-01

    Full Text Available Purpose : Aim of the COBRA (Consortium for Brachytherapy Data Analysis project is to create a multicenter group (consortium and a web-based system for standardized data collection. Material and methods: GEC-ESTRO (Groupe Européen de Curiethérapie – European Society for Radiotherapy & Oncology Head and Neck (H&N Working Group participated in the project and in the implementation of the consortium agreement, the ontology (data-set and the necessary COBRA software services as well as the peer reviewing of the general anatomic site-specific COBRA protocol. The ontology was defined by a multicenter task-group. Results : Eleven centers from 6 countries signed an agreement and the consortium approved the ontology. We identified 3 tiers for the data set: Registry (epidemiology analysis, Procedures (prediction models and DSS, and Research (radiomics. The COBRA-Storage System (C-SS is not time-consuming as, thanks to the use of “brokers”, data can be extracted directly from the single center’s storage systems through a connection with “structured query language database” (SQL-DB, Microsoft Access®, FileMaker Pro®, or Microsoft Excel®. The system is also structured to perform automatic archiving directly from the treatment planning system or afterloading machine. The architecture is based on the concept of “on-purpose data projection”. The C-SS architecture is privacy protecting because it will never make visible data that could identify an individual patient. This C-SS can also benefit from the so called “distributed learning” approaches, in which data never leave the collecting institution, while learning algorithms and proposed predictive models are commonly shared. Conclusions : Setting up a consortium is a feasible and practicable tool in the creation of an international and multi-system data sharing system. COBRA C-SS seems to be well accepted by all involved parties, primarily because it does not influence the center’s own

  9. A Web-Based Tool for Automatic Data Collection, Curation, and Visualization of Complex Healthcare Survey Studies including Social Network Analysis

    Directory of Open Access Journals (Sweden)

    José Alberto Benítez

    2017-01-01

    Full Text Available There is a great concern nowadays regarding alcohol consumption and drug abuse, especially in young people. Analyzing the social environment where these adolescents are immersed, as well as a series of measures determining the alcohol abuse risk or personal situation and perception using a number of questionnaires like AUDIT, FAS, KIDSCREEN, and others, it is possible to gain insight into the current situation of a given individual regarding his/her consumption behavior. But this analysis, in order to be achieved, requires the use of tools that can ease the process of questionnaire creation, data gathering, curation and representation, and later analysis and visualization to the user. This research presents the design and construction of a web-based platform able to facilitate each of the mentioned processes by integrating the different phases into an intuitive system with a graphical user interface that hides the complexity underlying each of the questionnaires and techniques used and presenting the results in a flexible and visual way, avoiding any manual handling of data during the process. Advantages of this approach are shown and compared to the previous situation where some of the tasks were accomplished by time consuming and error prone manipulations of data.

  10. A Web-Based Tool for Automatic Data Collection, Curation, and Visualization of Complex Healthcare Survey Studies including Social Network Analysis.

    Science.gov (United States)

    Benítez, José Alberto; Labra, José Emilio; Quiroga, Enedina; Martín, Vicente; García, Isaías; Marqués-Sánchez, Pilar; Benavides, Carmen

    2017-01-01

    There is a great concern nowadays regarding alcohol consumption and drug abuse, especially in young people. Analyzing the social environment where these adolescents are immersed, as well as a series of measures determining the alcohol abuse risk or personal situation and perception using a number of questionnaires like AUDIT, FAS, KIDSCREEN, and others, it is possible to gain insight into the current situation of a given individual regarding his/her consumption behavior. But this analysis, in order to be achieved, requires the use of tools that can ease the process of questionnaire creation, data gathering, curation and representation, and later analysis and visualization to the user. This research presents the design and construction of a web-based platform able to facilitate each of the mentioned processes by integrating the different phases into an intuitive system with a graphical user interface that hides the complexity underlying each of the questionnaires and techniques used and presenting the results in a flexible and visual way, avoiding any manual handling of data during the process. Advantages of this approach are shown and compared to the previous situation where some of the tasks were accomplished by time consuming and error prone manipulations of data.

  11. Automatic detection of adverse events to predict drug label changes using text and data mining techniques.

    Science.gov (United States)

    Gurulingappa, Harsha; Toldo, Luca; Rajput, Abdul Mateen; Kors, Jan A; Taweel, Adel; Tayrouz, Yorki

    2013-11-01

    The aim of this study was to assess the impact of automatically detected adverse event signals from text and open-source data on the prediction of drug label changes. Open-source adverse effect data were collected from FAERS, Yellow Cards and SIDER databases. A shallow linguistic relation extraction system (JSRE) was applied for extraction of adverse effects from MEDLINE case reports. Statistical approach was applied on the extracted datasets for signal detection and subsequent prediction of label changes issued for 29 drugs by the UK Regulatory Authority in 2009. 76% of drug label changes were automatically predicted. Out of these, 6% of drug label changes were detected only by text mining. JSRE enabled precise identification of four adverse drug events from MEDLINE that were undetectable otherwise. Changes in drug labels can be predicted automatically using data and text mining techniques. Text mining technology is mature and well-placed to support the pharmacovigilance tasks. Copyright © 2013 John Wiley & Sons, Ltd.

  12. Bohunice Simulator Data Collection Project

    International Nuclear Information System (INIS)

    Cillik, Ivan; Prochaska, Jan

    2002-01-01

    The paper describes the way and results of human reliability data analysis collected as a part of the Bohunice Simulator Data Collection Project (BSDCP), which was performed by VUJE Trnava, Inc. with funding support from the U.S. DOE, National Nuclear Security Administration. The goal of the project was to create a methodology for simulator data collection and analysis to support activities in probabilistic safety assessment (PSA) and human reliability assessment for Jaslovske Bohunice nuclear power plant consisting of two sets of twin units: two VVER 440/V-230 (V1) and two VVER 440/V-213 (V2) reactors. During the project training of V-2 control room crews was performed at VUJE-Trnava simulator. The simulator training and the data collection were done in parallel. The main goal of BSDCP was to collect suitable data of human errors under simulated conditions requiring the use of symptom-based emergency operating procedures (SBEOPs). The subjects of the data collection were scenario progress time data, operator errors, and real-time technological parameters. The paper contains three main parts. The first part presents preparatory work and semi-automatic computer-based methods used to collect data and to check technological parameters in order to find hidden errors of operators, to be able to retrace the course of each scenario for purposes of further analysis, and to document the whole training process. The first part gives also an overview of collected data scope, human error taxonomy, and state classifications for SBEOP instructions coding. The second part describes analytical work undertaken to describe time distribution necessary for execution of various kinds of instructions performed by operators according to the classification for coding of SBEOP instructions. It also presents the methods used for determination of probability distribution for different operator errors. Results from the data evaluation are presented in the last part of the paper. An overview of

  13. Bohunice Simulator Data Collection Project

    International Nuclear Information System (INIS)

    Cillik, I.; Prochaska, J.

    2002-01-01

    The paper describes the way and results of human reliability data analysis collected as a part of the Bohunice Simulator Data Collection Project (BSDCP), which was performed by VUJE Trnava, Inc. with funding support from the U.S. DOE, National Nuclear Security Administration. The goal of the project was to create a methodology for simulator data collection and analysis to support activities in probabilistic safety assessment (PSA) and human reliability assessment for Jaslovske Bohunice nuclear power plant consisting of two sets of twin units: two VVER 440/V-230 (V1) and two VVER 440/V-213 (V2) reactors. During the project, training of V-2 control room crews was performed at VUJE Trnava simulator. The simulator training and the data collection were done in parallel. The main goal of BSDCP was to collect suitable data of human errors under simulated conditions requiring the use of symptom-based emergency operating procedures (SBEOPs). The subjects of the data collection were scenario progress time data, operator errors, and real-time technological parameters. The paper contains three main parts. The first part presents preparatory work and semi-automatic computer-based methods used to collect data and to check technological parameters in order to find hidden errors of operators, to be able to retrace the course of each scenario for purposes of further analysis, and to document the whole training process. The first part gives also an overview of collected data scope, human error taxonomy, and state classifications for SBEOP instructions coding. The second part describes analytical work undertaken to describe time distribution necessary for execution of various kinds of instructions performed by operators according to the classification for coding of SBEOP instructions. It also presents the methods used for determination of probability distribution for various operator errors. Results from the data evaluation are presented in the last part of the paper. An overview of

  14. Electronic thermal sensor and Data Collection Platform technology: Part 5 in Thermal surveillance of active volcanoes using the Landsat-1 Data Collection System

    Science.gov (United States)

    Preble, Duane M.; Friedman, Jules D.; Frank, David

    1976-01-01

    Five Data Collection Platforms (DCP) were integrated electronically with thermall sensing systems, emplaced and operated in an analog mode at selected thermally significant volcanic and geothermal sites. The DCP's transmitted 3260 messages comprising 26,080 ambient, surface, and near-surface temperature records at an accuracy of ±1.15 °C for 1121 instrument days between November 14, 1972 and April 17, 1974. In harsh, windy, high-altitude volcanic environments the DCP functioned best with a small dipole antenna. Sixteen kg of alkaline batteries provided a viable power supply for the DCP systems, operated at a low-duty cycle, for 5 to 8 months. A proposed solar power supply system would lengthen the period of unattended operation of the system considerably. Special methods of data handling such as data storage via a proposed memory system would increase the significance of the twice-daily data reception enabling the DCP's to record full diurnal-temperature cycles at volcanic or geothermal sites. Refinements in the temperature-monitoring system designed and operated in experiment SR 251 included a backup system consisting of a multipoint temperature scanner, a servo mechanism and an analog-to-digital recorder. Improvements were made in temperature-probe design and in construction of corrosion-resistant seals by use of a hydrofluoric-acid-etching technique.

  15. Research on wireless communication technology based on automatic logistics system of welder

    Directory of Open Access Journals (Sweden)

    Sun Xuan

    2018-01-01

    Full Text Available In order to meet the requirements of high real-time and high stability of data transmission in automatic welding system, RTU data format and real-time communication mechanism are adopted in this system. In the automatic logistics system through the Ethernet and wireless WIFI technology will palletizer, stacker, AGV car organically together to complete the palletizer automatic crawling the goods, AGV car automatic delivery, stacking machine automatically out of the Dimensional warehouse. .

  16. Research on wireless communication technology based on automatic logistics system of welder

    OpenAIRE

    Sun Xuan; Wang Zhi-yong; Ma Zhe-dong

    2018-01-01

    In order to meet the requirements of high real-time and high stability of data transmission in automatic welding system, RTU data format and real-time communication mechanism are adopted in this system. In the automatic logistics system through the Ethernet and wireless WIFI technology will palletizer, stacker, AGV car organically together to complete the palletizer automatic crawling the goods, AGV car automatic delivery, stacking machine automatically out of the Dimensional warehouse. .

  17. Automatized system of radioactive material analysis

    International Nuclear Information System (INIS)

    Pchelkin, V.A.; Sviderskij, M.F.; Litvinov, V.A.; Lavrikov, S.A.

    1979-01-01

    An automatized system has been developed for the identification of substance, element and isotope content of radioactive materials on the basis of data obtained for studying physical-chemical properties of substances (with the help of atomic-absorption spectrometers, infrared spectrometer, mass-spectrometer, derivatograph etc.). The system is based on the following principles: independent operation of each device; a possibility of increasing the number of physical instruments and devices; modular properties of engineering and computer means; modular properties and standardization of mathematical equipment, high reliability of the system; continuity of programming languages; a possibility of controlling the devices with the help of high-level language, typification of the system; simple and easy service; low cost. Block-diagram of the system is given

  18. Upgradation of automatic liquid scintillation counting system

    International Nuclear Information System (INIS)

    Bhattacharya, Sadhana; Behere, Anita; Sonalkar, S.Y.; Vaidya, P.P.

    2001-01-01

    This paper describes the upgradation of Microprocessor based Automatic Liquid Scintillation Counting systems (MLSC). This system was developed in 1980's and subsequently many systems were manufactured and supplied to Environment Survey labs at various Nuclear Power Plants. Recently this system has been upgraded to a more sophisticated one by using PC add-on hardware and developing Windows based software. The software implements more intuitive graphical user interface and also enhances the features making it comparable with commercially available systems. It implements data processing using full spectrum analysis as against channel ratio method adopted earlier, improving the accuracy of the results. Also it facilitates qualitative as well as quantitative analysis of the β-spectrum. It is possible to analyze a sample containing an unknown β-source. (author)

  19. Using global positioning systems in health research a practical approach to data collection and processing

    DEFF Research Database (Denmark)

    Kerr, Jacqueline; Duncan, Scott; Schipperijn, Jasper

    2011-01-01

    The use of GPS devices in health research is increasingly popular. There are currently no best-practice guidelines for collecting, processing, and analyzing GPS data. The standardization of data collection and processing procedures will improve data quality, allow more-meaningful comparisons acro...

  20. Georeferenced and secure mobile health system for large scale data collection in primary care.

    Science.gov (United States)

    Sa, Joao H G; Rebelo, Marina S; Brentani, Alexandra; Grisi, Sandra J F E; Iwaya, Leonardo H; Simplicio, Marcos A; Carvalho, Tereza C M B; Gutierrez, Marco A

    2016-10-01

    Mobile health consists in applying mobile devices and communication capabilities for expanding the coverage and improving the effectiveness of health care programs. The technology is particularly promising for developing countries, in which health authorities can take advantage of the flourishing mobile market to provide adequate health care to underprivileged communities, especially primary care. In Brazil, the Primary Care Information System (SIAB) receives primary health care data from all regions of the country, creating a rich database for health-related action planning. Family Health Teams (FHTs) collect this data in periodic visits to families enrolled in governmental programs, following an acquisition procedure that involves filling in paper forms. This procedure compromises the quality of the data provided to health care authorities and slows down the decision-making process. To develop a mobile system (GeoHealth) that should address and overcome the aforementioned problems and deploy the proposed solution in a wide underprivileged metropolitan area of a major city in Brazil. The proposed solution comprises three main components: (a) an Application Server, with a database containing family health conditions; and two clients, (b) a Web Browser running visualization tools for management tasks, and (c) a data-gathering device (smartphone) to register and to georeference the family health data. A data security framework was designed to ensure the security of data, which was stored locally and transmitted over public networks. The system was successfully deployed at six primary care units in the city of Sao Paulo, where a total of 28,324 families/96,061 inhabitants are regularly followed up by government health policies. The health conditions observed from the population covered were: diabetes in 3.40%, hypertension (age >40) in 23.87% and tuberculosis in 0.06%. This estimated prevalence has enabled FHTs to set clinical appointments proactively, with the aim of

  1. Usability of Low-Cost Android Data Collection System for Community-Based Participatory Research.

    Science.gov (United States)

    Salihu, Hamisu M; Salinas-Miranda, Abraham; Turner, DeAnne; King, Lindsey; Paothong, Arnut; Austin, Deborah; Berry, Estrellita Lo

    2016-01-01

    Android tablet computers can be valuable tools for data collection, but their usability has not been evaluated in community-based participatory research (CBPR). This article examines the usability of a low-cost bilingual touchscreen computerized survey system using Android tablets, piloted with a sample of 201 community residents in Tampa, Florida, from November 2013 to March 2014. Needs assessment questions were designed with the droidSURVEY software, and deployed using Android tablet computers. In addition, participants were asked questions about system usability. The mean system usability was 77.57 ± 17.66 (range, 0-100). The mean completion time for taking the 63 survey questions in the needs assessment was 23.11 ± 9.62 minutes. The survey completion rate was optimal (100%), with only 6.34% missingness per variable. We found no sociodemographic differences in usability scores. Our findings indicate that Android tablets could serve as useful tools in CBPR studies.

  2. Artefact in Physiological Data Collected from Patients with Brain Injury: Quantifying the Problem and Providing a Solution Using a Factorial Switching Linear Dynamical Systems Approach.

    Science.gov (United States)

    Georgatzis, Konstantinos; Lal, Partha; Hawthorne, Christopher; Shaw, Martin; Piper, Ian; Tarbert, Claire; Donald, Rob; Williams, Christopher K I

    2016-01-01

    High-resolution, artefact-free and accurately annotated physiological data are desirable in patients with brain injury both to inform clinical decision-making and for intelligent analysis of the data in applications such as predictive modelling. We have quantified the quality of annotation surrounding artefactual events and propose a factorial switching linear dynamical systems (FSLDS) approach to automatically detect artefact in physiological data collected in the neurological intensive care unit (NICU). Retrospective analysis of the BrainIT data set to discover potential hypotensive events corrupted by artefact and identify the annotation of associated clinical interventions. Training of an FSLDS model on clinician-annotated artefactual events in five patients with severe traumatic brain injury. In a subset of 187 patients in the BrainIT database, 26.5 % of potential hypotensive events were abandoned because of artefactual data. Only 30 % of these episodes could be attributed to an annotated clinical intervention. As assessed by the area under the receiver operating characteristic curve metric, FSLDS model performance in automatically identifying the events of blood sampling, arterial line damping and patient handling was 0.978, 0.987 and 0.765, respectively. The influence of artefact on physiological data collected in the NICU is a significant problem. This pilot study using an FSLDS approach shows real promise and is under further development.

  3. Data collection on component malfunctions and failures of JET ICRH system

    International Nuclear Information System (INIS)

    Pinna, T.; Cambi, G.

    2007-01-01

    The objective of the activity was to collect and analyse data coming out from operating experiences gained in the Joint European Torus (JET) for the Ion Cyclotron Resonance Heating (ICRH) system in order to enrich the data collection on failures of components used in fusion facilities. Alarms/Failures and malfunctions occurred in the years of operations from March 1996 to November 2005, including information on failure modes and, where possible, causes of the failures, have been identified. Beyond information on failures and alarms events, also data related to crowbar events have been collected. About 3400 events classified as alarms or failures related to specific components or sub-systems were identified by analysing the 25 hand-written logbooks made available by the ICRH operation staff. Information about the JET pulses in which the ICRH system was operated has been extracted from the tick sheets covering the whole considered time interval. 20 hand written tick sheets cover the period from March 1996 to middle May 2003, while tick sheets recorded as excel files cover the period from May 2003 to November 2005. By analysing the tick sheets it results that the ICRH was operated during about 12000 plasma pulses. Main statistical values, such as rates of alarms/failures and corresponding standard errors and confidence intervals, have been estimated. Failure rates of systems and components have been evaluated both with regard to the ICRH operation pulses and operating days (days in which at least one ICRH module was requested to operate). Failure probabilities on demand have been evaluated with regard to number of pulses operated. Some of the results are the following: - The highest number of alarms/failures (1243) appears to be related to Erratic /No-output of the Instrumentation and Control (I and C) apparatus, followed by faults (829) of the Tetrode circuits, by faults (466) of the High Voltage Power Supply system and by faults (428) of the Tuning elements. - The

  4. URBAN DATA COLLECTION USING A BIKE MOBILE SYSTEM WITH A FOSS ARCHITECTURE

    Directory of Open Access Journals (Sweden)

    A. Abdul Jabbar

    2017-07-01

    Full Text Available European community is working to improve the quality of the life in each European country, in particular to increase the quality air condition and safety in each city. The quality air is daily monitored, using several ground station, which do not consider the variation of the quality during the day, evaluating only the average level. In this case, it could be interesting to have a “smart” system to acquire distributed data in continuous, even involving the citizens. On the other hand, to improve the safety level in urban area along cycle lane, road and pedestrian path, exist a lot of algorithms for visibility and safety analysis; the crucial aspect is the 3D model considered as “input” in these algorithms, which always needs to be updated. A bike has been instrumented with two digital camera as Raspberry PI-cam. Image acquisition has been realized with a dedicated python tool, which has been implemented in the Raspberry PI system. Images have been georeferenced using a u-blox 8T, connected to Raspberry system. GNSS data has been acquired using a specific tool developed in Python, which was based on RTKLIB library. Time synchronization has been obtained with GNSS receiver. Additionally, a portable laser scanner, an air quality system and a small Inertial platform have been installed and connected with the Raspberry system. The system has been implemented and tested to acquire data (image and air quality parameter in a district in Turin. Also a 3D model of the investigated site has been carried. In this contribute, the assembling of the system is described, in particular the dataset acquired and the results carried out will be described. different low cost sensors, in particular digital camera and laser scanner to collect easily geospatial data in urban area.

  5. VetCompass Australia: A National Big Data Collection System for Veterinary Science

    Science.gov (United States)

    McGreevy, Paul; Thomson, Peter; Dhand, Navneet K.; Raubenheimer, David; Masters, Sophie; Mansfield, Caroline S.; Baldwin, Timothy; Soares Magalhaes, Ricardo J.; Rand, Jacquie; Hill, Peter; Gilkerson, James; Combs, Martin; Raidal, Shane; Irwin, Peter; Irons, Peter; Squires, Richard; Brodbelt, David; Hammond, Jeremy

    2017-01-01

    Simple Summary The VetCompass Australia program collects real-time clinical records from veterinary practices and aggregates them for researchers to interrogate. It delivers Australian researchers sustainable and cost-effective access to authoritative data from hundreds of veterinary practitioners, across Australia and opens up major international collaborative opportunities with related projects in the United Kingdom and elsewhere. Abstract VetCompass Australia is veterinary medical records-based research coordinated with the global VetCompass endeavor to maximize its quality and effectiveness for Australian companion animals (cats, dogs, and horses). Bringing together all seven Australian veterinary schools, it is the first nationwide surveillance system collating clinical records on companion-animal diseases and treatments. VetCompass data service collects and aggregates real-time, clinical records for researchers to interrogate, delivering sustainable and cost-effective access to data from hundreds of veterinary practitioners nationwide. Analysis of these clinical records will reveal geographical and temporal trends in the prevalence of inherited and acquired diseases, identify frequently prescribed treatments, revolutionize clinical auditing, help the veterinary profession to rank research priorities, and assure evidence-based companion-animal curricula in veterinary schools. VetCompass Australia will progress in three phases: (1) roll-out of the VetCompass platform to harvest Australian veterinary clinical record data; (2) development and enrichment of the coding (data-presentation) platform; and (3) creation of a world-first, real-time surveillance interface with natural language processing (NLP) technology. The first of these three phases is described in the current article. Advances in the collection and sharing of records from numerous practices will enable veterinary professionals to deliver a vastly improved level of care for companion animals that will

  6. VetCompass Australia: A National Big Data Collection System for Veterinary Science.

    Science.gov (United States)

    McGreevy, Paul; Thomson, Peter; Dhand, Navneet K; Raubenheimer, David; Masters, Sophie; Mansfield, Caroline S; Baldwin, Timothy; Soares Magalhaes, Ricardo J; Rand, Jacquie; Hill, Peter; Peaston, Anne; Gilkerson, James; Combs, Martin; Raidal, Shane; Irwin, Peter; Irons, Peter; Squires, Richard; Brodbelt, David; Hammond, Jeremy

    2017-09-26

    VetCompass Australia is veterinary medical records-based research coordinated with the global VetCompass endeavor to maximize its quality and effectiveness for Australian companion animals (cats, dogs, and horses). Bringing together all seven Australian veterinary schools, it is the first nationwide surveillance system collating clinical records on companion-animal diseases and treatments. VetCompass data service collects and aggregates real-time, clinical records for researchers to interrogate, delivering sustainable and cost-effective access to data from hundreds of veterinary practitioners nationwide. Analysis of these clinical records will reveal geographical and temporal trends in the prevalence of inherited and acquired diseases, identify frequently prescribed treatments, revolutionize clinical auditing, help the veterinary profession to rank research priorities, and assure evidence-based companion-animal curricula in veterinary schools. VetCompass Australia will progress in three phases: (1) roll-out of the VetCompass platform to harvest Australian veterinary clinical record data; (2) development and enrichment of the coding (data-presentation) platform; and (3) creation of a world-first, real-time surveillance interface with natural language processing (NLP) technology. The first of these three phases is described in the current article. Advances in the collection and sharing of records from numerous practices will enable veterinary professionals to deliver a vastly improved level of care for companion animals that will improve their quality of life.

  7. VetCompass Australia: A National Big Data Collection System for Veterinary Science

    Directory of Open Access Journals (Sweden)

    Paul McGreevy

    2017-09-01

    Full Text Available VetCompass Australia is veterinary medical records-based research coordinated with the global VetCompass endeavor to maximize its quality and effectiveness for Australian companion animals (cats, dogs, and horses. Bringing together all seven Australian veterinary schools, it is the first nationwide surveillance system collating clinical records on companion-animal diseases and treatments. VetCompass data service collects and aggregates real-time, clinical records for researchers to interrogate, delivering sustainable and cost-effective access to data from hundreds of veterinary practitioners nationwide. Analysis of these clinical records will reveal geographical and temporal trends in the prevalence of inherited and acquired diseases, identify frequently prescribed treatments, revolutionize clinical auditing, help the veterinary profession to rank research priorities, and assure evidence-based companion-animal curricula in veterinary schools. VetCompass Australia will progress in three phases: (1 roll-out of the VetCompass platform to harvest Australian veterinary clinical record data; (2 development and enrichment of the coding (data-presentation platform; and (3 creation of a world-first, real-time surveillance interface with natural language processing (NLP technology. The first of these three phases is described in the current article. Advances in the collection and sharing of records from numerous practices will enable veterinary professionals to deliver a vastly improved level of care for companion animals that will improve their quality of life.

  8. Feasibility Study for Ballet E-Learning: Automatic Composition System for Ballet "Enchainement" with Online 3D Motion Data Archive

    Science.gov (United States)

    Umino, Bin; Longstaff, Jeffrey Scott; Soga, Asako

    2009-01-01

    This paper reports on "Web3D dance composer" for ballet e-learning. Elementary "petit allegro" ballet steps were enumerated in collaboration with ballet teachers, digitally acquired through 3D motion capture systems, and categorised into families and sub-families. Digital data was manipulated into virtual reality modelling language (VRML) and fit…

  9. Development of automatic ultrasonic testing system and its application

    International Nuclear Information System (INIS)

    Oh, Sang Hong; Matsuura, Toshihiko; Iwata, Ryusuke; Nakagawa, Michio; Horikawa, Kohsuke; Kim, You Chul

    1997-01-01

    The radiographic testing (RT) has been usually applied to a nondestructive testing, which is carried out on purpose to detect internal defects at welded joints of a penstock. In the case that RT could not be applied to, the ultrasonic testing (UT) was performed. UT was generally carried out by manual scanning and the inspections data were recorded by the inspector in a site. So, as a weak point, there was no objective inspection records correspond to films of RT. It was expected that the automatic ultrasonic testing system by which automatic scanning and automatic recording are possible was developed. In this respect, the automatic ultrasonic testing system was developed. Using newly developed the automatic ultrasonic testing system, test results to the circumferential welded joints of the penstock at a site were shown in this paper.

  10. A cost-effective traffic data collection system based on the iDEN mobile telecommunication network.

    Science.gov (United States)

    2008-10-01

    This report describes a cost-effective data collection system for Caltrans 170 traffic signal : controller. The data collection system is based on TCP/IP communication over existing : low-cost mobile communication networks and Motorola iDEN1 mobile...

  11. Creating an iPhone Application for Collecting Continuous ABC Data

    Science.gov (United States)

    Whiting, Seth W.; Dixon, Mark R.

    2012-01-01

    This paper provides an overview and task analysis for creating a continuous ABC data- collection application using Xcode on a Mac computer. Behavior analysts can program an ABC data collection system, complete with a customized list of target clients, antecedents, behaviors, and consequences to be recorded, and have the data automatically sent to…

  12. Automatic generation of combinatorial test data

    CERN Document Server

    Zhang, Jian; Ma, Feifei

    2014-01-01

    This book reviews the state-of-the-art in combinatorial testing, with particular emphasis on the automatic generation of test data. It describes the most commonly used approaches in this area - including algebraic construction, greedy methods, evolutionary computation, constraint solving and optimization - and explains major algorithms with examples. In addition, the book lists a number of test generation tools, as well as benchmarks and applications. Addressing a multidisciplinary topic, it will be of particular interest to researchers and professionals in the areas of software testing, combi

  13. Programs for the automatic gamma-ray measurement with CANBERRA 8100/QUANTA system

    International Nuclear Information System (INIS)

    Yoshida, Hiroshi; Sakai, Eiji; Kubo, Katsumi.

    1982-07-01

    Some programs have been prepared for the automatic operation of the CANBERRA 8100/QUANTA System for the gamma-ray spectrum measurement. The main parts of these programs are: (1) to collect and record on magnetic disks the data of gamma-ray spectra automatically, while the recorded data are analyzed to estimate the nuclides which generate photopeaks of spectra and to calculate those concentrations; (2) to draw plotted diagrams of pulse height distributions of gamma-ray spectra data and other data by the additional digital plotter; and etc. (author)

  14. Fully automatized renal parenchyma volumetry using a support vector machine based recognition system for subject-specific probability map generation in native MR volume data

    Science.gov (United States)

    Gloger, Oliver; Tönnies, Klaus; Mensel, Birger; Völzke, Henry

    2015-11-01

    In epidemiological studies as well as in clinical practice the amount of produced medical image data strongly increased in the last decade. In this context organ segmentation in MR volume data gained increasing attention for medical applications. Especially in large-scale population-based studies organ volumetry is highly relevant requiring exact organ segmentation. Since manual segmentation is time-consuming and prone to reader variability, large-scale studies need automatized methods to perform organ segmentation. Fully automatic organ segmentation in native MR image data has proven to be a very challenging task. Imaging artifacts as well as inter- and intrasubject MR-intensity differences complicate the application of supervised learning strategies. Thus, we propose a modularized framework of a two-stepped probabilistic approach that generates subject-specific probability maps for renal parenchyma tissue, which are refined subsequently by using several, extended segmentation strategies. We present a three class-based support vector machine recognition system that incorporates Fourier descriptors as shape features to recognize and segment characteristic parenchyma parts. Probabilistic methods use the segmented characteristic parenchyma parts to generate high quality subject-specific parenchyma probability maps. Several refinement strategies including a final shape-based 3D level set segmentation technique are used in subsequent processing modules to segment renal parenchyma. Furthermore, our framework recognizes and excludes renal cysts from parenchymal volume, which is important to analyze renal functions. Volume errors and Dice coefficients show that our presented framework outperforms existing approaches.

  15. Fully automatized renal parenchyma volumetry using a support vector machine based recognition system for subject-specific probability map generation in native MR volume data

    International Nuclear Information System (INIS)

    Gloger, Oliver; Völzke, Henry; Tönnies, Klaus; Mensel, Birger

    2015-01-01

    In epidemiological studies as well as in clinical practice the amount of produced medical image data strongly increased in the last decade. In this context organ segmentation in MR volume data gained increasing attention for medical applications. Especially in large-scale population-based studies organ volumetry is highly relevant requiring exact organ segmentation. Since manual segmentation is time-consuming and prone to reader variability, large-scale studies need automatized methods to perform organ segmentation. Fully automatic organ segmentation in native MR image data has proven to be a very challenging task. Imaging artifacts as well as inter- and intrasubject MR-intensity differences complicate the application of supervised learning strategies. Thus, we propose a modularized framework of a two-stepped probabilistic approach that generates subject-specific probability maps for renal parenchyma tissue, which are refined subsequently by using several, extended segmentation strategies. We present a three class-based support vector machine recognition system that incorporates Fourier descriptors as shape features to recognize and segment characteristic parenchyma parts. Probabilistic methods use the segmented characteristic parenchyma parts to generate high quality subject-specific parenchyma probability maps. Several refinement strategies including a final shape-based 3D level set segmentation technique are used in subsequent processing modules to segment renal parenchyma. Furthermore, our framework recognizes and excludes renal cysts from parenchymal volume, which is important to analyze renal functions. Volume errors and Dice coefficients show that our presented framework outperforms existing approaches. (paper)

  16. Development of advanced automatic operation system for nuclear ship. 1. Perfect automatic normal operation

    International Nuclear Information System (INIS)

    Nakazawa, Toshio; Yabuuti, Noriaki; Takahashi, Hiroki; Shimazaki, Junya

    1999-02-01

    Development of operation support system such as automatic operating system and anomaly diagnosis systems of nuclear reactor is very important in practical nuclear ship because of a limited number of operators and severe conditions in which receiving support from others in a case of accident is very difficult. The goal of development of the operation support systems is to realize the perfect automatic control system in a series of normal operation from the reactor start-up to the shutdown. The automatic control system for the normal operation has been developed based on operating experiences of the first Japanese nuclear ship 'Mutsu'. Automation technique was verified by 'Mutsu' plant data at manual operation. Fully automatic control of start-up and shutdown operations was achieved by setting the desired value of operation and the limiting value of parameter fluctuation, and by making the operation program of the principal equipment such as the main coolant pump and the heaters. This report presents the automatic operation system developed for the start-up and the shutdown of reactor and the verification of the system using the Nuclear Ship Engineering Simulator System. (author)

  17. Data collection system for a wide range of gas-discharge proportional neutron counters

    Science.gov (United States)

    Oskomov, V.; Sedov, A.; Saduyev, N.; Kalikulov, O.; Kenzhina, I.; Tautaev, E.; Mukhamejanov, Y.; Dyachkov, V.; Utey, Sh

    2017-12-01

    This article describes the development and creation of a universal system of data collection to measure the intensity of pulsed signals. As a result of careful analysis of time conditions and operating conditions of software and hardware complex circuit solutions were selected that meet the required specifications: frequency response is optimized in order to obtain the maximum ratio signal/noise; methods and modes of operation of the microcontroller were worked out to implement the objectives of continuous measurement of signal amplitude at the output of amplifier and send the data to a computer; function of control of high voltage source was implemented. The preliminary program has been developed for microcontroller in its simplest form, which works on a particular algorithm.

  18. Automatic processing of radioimmunological research data on a computer

    International Nuclear Information System (INIS)

    Korolyuk, I.P.; Gorodenko, A.N.; Gorodenko, S.I.

    1979-01-01

    A program ''CRITEST'' in the language PL/1 for the EC computer intended for automatic processing of the results of radioimmunological research has been elaborated. The program works in the operation system of the OC EC computer and is performed in the section OC 60 kb. When compiling the program Eitken's modified algorithm was used. The program was clinically approved when determining a number of hormones: CTH, T 4 , T 3 , TSH. The automatic processing of the radioimmunological research data on the computer makes it possible to simplify the labour-consuming analysis and to raise its accuracy

  19. submitter Optimizing the data-collection time of a large-scale data-acquisition system through a simulation framework

    CERN Document Server

    Colombo, Tommaso; Garcìa, Pedro Javier; Vandelli, Wainer

    2016-01-01

    The ATLAS detector at CERN records particle collision “events” delivered by the Large Hadron Collider. Its data-acquisition system identifies, selects, and stores interesting events in near real-time, with an aggregate throughput of several 10 GB/s. It is a distributed software system executed on a farm of roughly 2000 commodity worker nodes communicating via TCP/IP on an Ethernet network. Event data fragments are received from the many detector readout channels and are buffered, collected together, analyzed and either stored permanently or discarded. This system, and data-acquisition systems in general, are sensitive to the latency of the data transfer from the readout buffers to the worker nodes. Challenges affecting this transfer include the many-to-one communication pattern and the inherently bursty nature of the traffic. The main performance issues brought about by this workload are addressed in this paper, focusing in particular on the so-called TCP incast pathology. Since performing systematic stud...

  20. WP 2: "Data collection and processing systems (DCPS) for the conventional markets" and WP 3: "Data collection and processing systems for organic markets"

    NARCIS (Netherlands)

    Wolfert, J.; Kramer, K.J.; Richter, T.; Hempfling, G.; Lux, S.; Recke, G.

    2004-01-01

    The aim of the EU concerted action EISfOM (QLK5-2002-02400) (European Information System for Organic Markets) is to build up a framework for reporting valid and reliable data for relevant production and market sectors of the European organic sector in order to meet the needs of policy-makers,

  1. Steam System Balancing and Tuning for Multifamily Residential Buildings in Chicagoland - Second Year of Data Collection

    Energy Technology Data Exchange (ETDEWEB)

    Choi, J.; Ludwig, P.; Brand, L.

    2013-08-01

    Steam heated buildings often suffer from uneven heating as a result of poor control of the amount of steam entering each radiator. In order to satisfy the heating load to the coldest units, other units are overheated. As a result, some tenants complain of being too hot and open their windows in the middle of winter, while others complain of being too cold and are compelled to use supplemental heat sources. Building on previous research, CNT Energy identified 10 test buildings in Chicago and conducted a study to identify best practices for the methodology, typical costs, and energy savings associated with steam system balancing. A package of common steam balancing measures was assembled and data were collected on the buildings before and after these retrofits were installed to investigate the process, challenges, and the cost effectiveness of improving steam systems through improved venting and control systems. The test buildings that received venting upgrades and new control systems showed 10.2% savings on their natural gas heating load, with a simple payback of 5.1 years. The methodologies for and findings from this study are presented in detail in this report. This report has been updated from a version published in August 2012 to include natural gas usage information from the 2012 heating season and updated natural gas savings calculations.

  2. CCTV Data Management for Effective Condition Assessment of Wastewater Collection Systems - slides

    Science.gov (United States)

    The project objectives are: Research current state of condition assessment technology for wastewater collection systems; Evaluate performance and cost of innovative and advanced infrastructure monitoring technologies; Identify and evaluate innovative CCTV technologies; and Prepar...

  3. Get SMARTS] (Sports Medicine Research Team System): A Computerized Outpatient Data Collection System for Epidemiologic Research

    National Research Council Canada - National Science Library

    Brodine, S

    1997-01-01

    .... This report describes features of the Sports Medicine Research Team System (SMARTS) and reviews results of a SMARTS supported prospective study of male Marine Corps recruits undergoing basic training...

  4. Using global positioning systems in health research: a practical approach to data collection and processing.

    Science.gov (United States)

    Kerr, Jacqueline; Duncan, Scott; Schipperijn, Jasper; Schipperjin, Jasper

    2011-11-01

    The use of GPS devices in health research is increasingly popular. There are currently no best-practice guidelines for collecting, processing, and analyzing GPS data. The standardization of data collection and processing procedures will improve data quality, allow more-meaningful comparisons across studies and populations, and advance this field more rapidly. This paper aims to take researchers, who are considering using GPS devices in their research, through device-selection criteria, device settings, participant data collection, data cleaning, data processing, and integration of data into GIS. Recommendations are outlined for each stage of data collection and analysis and indicates challenges that should be considered. This paper highlights the benefits of collecting GPS data over traditional self-report or estimated exposure measures. Information presented here will allow researchers to make an informed decision about incorporating this readily available technology into their studies. This work reflects the state of the art in 2011. Copyright © 2011 American Journal of Preventive Medicine. Published by Elsevier Inc. All rights reserved.

  5. DataCollection Prototyping

    CERN Multimedia

    Beck, H.P.

    DataCollection is a subsystem of the Trigger, DAQ & DCS project responsible for the movement of event data from the ROS to the High Level Triggers. This includes data from Regions of Interest (RoIs) for Level 2, building complete events for the Event Filter and finally transferring accepted events to Mass Storage. It also handles passing the LVL1 RoI pointers and the allocation of Level 2 processors and load balancing of Event Building. During the last 18 months DataCollection has developed a common architecture for the hardware and software required. This involved a radical redesign integrating ideas from separate parts of earlier TDAQ work. An important milestone for this work, now achieved, has been to demonstrate this subsystem in the so-called Phase 2A Integrated Prototype. This prototype comprises the various TDAQ hardware and software components (ROSs, LVL2, etc.) under the control of the TDAQ Online software. The basic functionality has been demonstrated on small testbeds (~8-10 processing nodes)...

  6. The MPO system for automatic workflow documentation

    Energy Technology Data Exchange (ETDEWEB)

    Abla, G.; Coviello, E.N.; Flanagan, S.M. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Greenwald, M. [Massachusetts Institute of Technology, Cambridge, MA 02139 (United States); Lee, X. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Romosan, A. [Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); Schissel, D.P., E-mail: schissel@fusion.gat.com [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Shoshani, A. [Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); Stillerman, J.; Wright, J. [Massachusetts Institute of Technology, Cambridge, MA 02139 (United States); Wu, K.J. [Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States)

    2016-11-15

    Highlights: • Data model, infrastructure, and tools for data tracking, cataloging, and integration. • Automatically document workflow and data provenance in the widest sense. • Fusion Science as test bed but the system’s framework and data model is quite general. - Abstract: Data from large-scale experiments and extreme-scale computing is expensive to produce and may be used for critical applications. However, it is not the mere existence of data that is important, but our ability to make use of it. Experience has shown that when metadata is better organized and more complete, the underlying data becomes more useful. Traditionally, capturing the steps of scientific workflows and metadata was the role of the lab notebook, but the digital era has resulted instead in the fragmentation of data, processing, and annotation. This paper presents the Metadata, Provenance, and Ontology (MPO) System, the software that can automate the documentation of scientific workflows and associated information. Based on recorded metadata, it provides explicit information about the relationships among the elements of workflows in notebook form augmented with directed acyclic graphs. A set of web-based graphical navigation tools and Application Programming Interface (API) have been created for searching and browsing, as well as programmatically accessing the workflows and data. We describe the MPO concepts and its software architecture. We also report the current status of the software as well as the initial deployment experience.

  7. The MPO system for automatic workflow documentation

    International Nuclear Information System (INIS)

    Abla, G.; Coviello, E.N.; Flanagan, S.M.; Greenwald, M.; Lee, X.; Romosan, A.; Schissel, D.P.; Shoshani, A.; Stillerman, J.; Wright, J.; Wu, K.J.

    2016-01-01

    Highlights: • Data model, infrastructure, and tools for data tracking, cataloging, and integration. • Automatically document workflow and data provenance in the widest sense. • Fusion Science as test bed but the system’s framework and data model is quite general. - Abstract: Data from large-scale experiments and extreme-scale computing is expensive to produce and may be used for critical applications. However, it is not the mere existence of data that is important, but our ability to make use of it. Experience has shown that when metadata is better organized and more complete, the underlying data becomes more useful. Traditionally, capturing the steps of scientific workflows and metadata was the role of the lab notebook, but the digital era has resulted instead in the fragmentation of data, processing, and annotation. This paper presents the Metadata, Provenance, and Ontology (MPO) System, the software that can automate the documentation of scientific workflows and associated information. Based on recorded metadata, it provides explicit information about the relationships among the elements of workflows in notebook form augmented with directed acyclic graphs. A set of web-based graphical navigation tools and Application Programming Interface (API) have been created for searching and browsing, as well as programmatically accessing the workflows and data. We describe the MPO concepts and its software architecture. We also report the current status of the software as well as the initial deployment experience.

  8. An observing system for the collection of fishery and oceanographic data

    Directory of Open Access Journals (Sweden)

    P. Falco

    2007-05-01

    Full Text Available Fishery Observing System (FOS was developed as a first and basic step towards fish stock abundance nowcasting/forecasting within the framework of the EU research program Mediterranean Forecasting System: Toward an Environmental Prediction (MFSTEP. The study of the relationship between abundance and environmental parameters also represents a crucial point towards forecasting. Eight fishing vessels were progressively equipped with FOS instrumentation to collect fishery and oceanographic data. The vessels belonged to different harbours of the Central and Northern Adriatic Sea. For this pilot application, anchovy (Engraulis encrasicolus, L. was chosen as the target species. Geo-referenced catch data, associated with in-situ temperature and depth, were the FOS products but other parameters were associated with catch data as well. MFSTEP numerical circulation models provide many of these data. In particular, salinity was extracted from re-analysis data of numerical circulation models. Satellite-derived sea surface temperature (SST and chlorophyll were also used as independent variables. Catch and effort data were used to estimate an abundance index (CPUE – Catch per Unit of Effort. Considering that catch records were gathered by different fishing vessels with different technical characteristics and operating on different fish densities, a standardized value of CPUE was calculated. A spatial and temporal average CPUE map was obtained together with a monthly mean time series in order to characterise the variability of anchovy abundance during the period of observation (October 2003–August 2005. In order to study the relationship between abundance and oceanographic parameters, Generalized Additive Models (GAM were used. Preliminary results revealed a complex scenario: the southern sector of the domain is characterised by a stronger relationship than the central and northern sector where the interactions between the environment and the anchovy

  9. Automatic control variac system for electronic accelerator

    International Nuclear Information System (INIS)

    Zhang Shuocheng; Wang Dan; Jing Lan; Qiao Weimin; Ma Yunhai

    2006-01-01

    An automatic control variac system is designed in order to satisfy the controlling requirement of the electronic accelerator developed by the Institute. Both design and operational principles, structure of the system as well as the software of industrial PC and micro controller unit are described. The interfaces of the control module are RS232 and RS485. A fiber optical interface (FOC) could be set up if an industrial FOC network is necessary, which will extend the filed of its application and make the communication of the system better. It is shown in practice that the system can adjust the variac output voltage automatically and assure the accurate and automatic control of the electronic accelerator. The system is designed in accordance with the general design principles and possesses the merits such as easy operation and maintenance, good expansibility, and low cost, thus it could also be used in other industrial branches. (authors)

  10. Automatic weld torch guidance control system

    Science.gov (United States)

    Smaith, H. E.; Wall, W. A.; Burns, M. R., Jr.

    1982-01-01

    A highly reliable, fully digital, closed circuit television optical, type automatic weld seam tracking control system was developed. This automatic tracking equipment is used to reduce weld tooling costs and increase overall automatic welding reliability. The system utilizes a charge injection device digital camera which as 60,512 inidividual pixels as the light sensing elements. Through conventional scanning means, each pixel in the focal plane is sequentially scanned, the light level signal digitized, and an 8-bit word transmitted to scratch pad memory. From memory, the microprocessor performs an analysis of the digital signal and computes the tracking error. Lastly, the corrective signal is transmitted to a cross seam actuator digital drive motor controller to complete the closed loop, feedback, tracking system. This weld seam tracking control system is capable of a tracking accuracy of + or - 0.2 mm, or better. As configured, the system is applicable to square butt, V-groove, and lap joint weldments.

  11. Automatic dam concrete placing system; Dam concrete dasetsu sagyo no jidoka system

    Energy Technology Data Exchange (ETDEWEB)

    Yoneda, Y; Hori, Y; Nakayama, T; Yoshihara, K; Hironaka, T [Okumura Corp., Osaka (Japan)

    1994-11-15

    An automatic concrete placing system was developed for concrete dam construction. This system consists of the following five subsystems: a wireless data transmission system, an automatic dam concrete mixing system, a consistency determination system, an automatic dam concrete loading and transporting system, and a remote concrete bucket opening and closing system. The system includes the following features: mixing amount by mixing ratio and mixing intervals can be instructed from a concrete placing site by using a wireless handy terminal; concrete is mixed automatically in a batcher plant; a transfer car is started, and concrete is charged into a bucket automatically; the mixed concrete is determined of its properties automatically; labor cost can be reduced, the work efficiency improved, and the safety enhanced; and the system introduction has resulted in unattended operation from the aggregate draw-out to a bunker line, manpower saving of five persons, and reduction in cycle time by 10%. 11 figs., 2 tabs.

  12. Randomized algorithms in automatic control and data mining

    CERN Document Server

    Granichin, Oleg; Toledano-Kitai, Dvora

    2015-01-01

    In the fields of data mining and control, the huge amount of unstructured data and the presence of uncertainty in system descriptions have always been critical issues. The book Randomized Algorithms in Automatic Control and Data Mining introduces the readers to the fundamentals of randomized algorithm applications in data mining (especially clustering) and in automatic control synthesis. The methods proposed in this book guarantee that the computational complexity of classical algorithms and the conservativeness of standard robust control techniques will be reduced. It is shown that when a problem requires "brute force" in selecting among options, algorithms based on random selection of alternatives offer good results with certain probability for a restricted time and significantly reduce the volume of operations.

  13. Robust indexing for automatic data collection

    International Nuclear Information System (INIS)

    Sauter, Nicholas K.; Grosse-Kunstleve, Ralf W.; Adams, Paul D.

    2003-01-01

    We present improved methods for indexing diffraction patterns from macromolecular crystals. The novel procedures include a more robust way to verify the position of the incident X-ray beam on the detector, an algorithm to verify that the deduced lattice basis is consistent with the observations, and an alternative approach to identify the metric symmetry of the lattice. These methods help to correct failures commonly experienced during indexing, and increase the overall success rate of the process. Rapid indexing, without the need for visual inspection, will play an important role as beamlines at synchrotron sources prepare for high-throughput automation

  14. Development of automatic laser welding system

    International Nuclear Information System (INIS)

    Ohwaki, Katsura

    2002-01-01

    Laser are a new production tool for high speed and low distortion welding and applications to automatic welding lines are increasing. IHI has long experience of laser processing for the preservation of nuclear power plants, welding of airplane engines and so on. Moreover, YAG laser oscillators and various kinds of hardware have been developed for laser welding and automation. Combining these welding technologies and laser hardware technologies produce the automatic laser welding system. In this paper, the component technologies are described, including combined optics intended to improve welding stability, laser oscillators, monitoring system, seam tracking system and so on. (author)

  15. 14 CFR 23.1329 - Automatic pilot system.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 1 2010-01-01 2010-01-01 false Automatic pilot system. 23.1329 Section 23...: Installation § 23.1329 Automatic pilot system. If an automatic pilot system is installed, it must meet the following: (a) Each system must be designed so that the automatic pilot can— (1) Be quickly and positively...

  16. Implementation and flight tests for the Digital Integrated Automatic Landing System (DIALS). Part 1: Flight software equations, flight test description and selected flight test data

    Science.gov (United States)

    Hueschen, R. M.

    1986-01-01

    Five flight tests of the Digital Automated Landing System (DIALS) were conducted on the Advanced Transport Operating Systems (ATOPS) Transportation Research Vehicle (TSRV) -- a modified Boeing 737 aircraft for advanced controls and displays research. These flight tests were conducted at NASA's Wallops Flight Center using the microwave landing system (MLS) installation on runway 22. This report describes the flight software equations of the DIALS which was designed using modern control theory direct-digital design methods and employed a constant gain Kalman filter. Selected flight test performance data is presented for localizer (runway centerline) capture and track at various intercept angles, for glideslope capture and track of 3, 4.5, and 5 degree glideslopes, for the decrab maneuver, and for the flare maneuver. Data is also presented to illustrate the system performance in the presence of cross, gust, and shear winds. The mean and standard deviation of the peak position errors for localizer capture were, respectively, 24 feet and 26 feet. For mild wind conditions, glideslope and localizer tracking position errors did not exceed, respectively, 5 and 20 feet. For gusty wind conditions (8 to 10 knots), these errors were, respectively, 10 and 30 feet. Ten hands off automatic lands were performed. The standard deviation of the touchdown position and velocity errors from the mean values were, respectively, 244 feet and 0.7 feet/sec.

  17. Automatic seismic support design of piping system by an object oriented expert system

    International Nuclear Information System (INIS)

    Nakatogawa, T.; Takayama, Y.; Hayashi, Y.; Fukuda, T.; Yamamoto, Y.; Haruna, T.

    1990-01-01

    The seismic support design of piping systems of nuclear power plants requires many experienced engineers and plenty of man-hours, because the seismic design conditions are very severe, the bulk volume of the piping systems is hyge and the design procedures are very complicated. Therefore we have developed a piping seismic design expert system, which utilizes the piping design data base of a 3 dimensional CAD system and automatically determines the piping support locations and support styles. The data base of this system contains the maximum allowable seismic support span lengths for straight piping and the span length reduction factors for bends, branches, concentrated masses in the piping, and so forth. The system automatically produces the support design according to the design knowledge extracted and collected from expert design engineers, and using design information such as piping specifications which give diameters and thickness and piping geometric configurations. The automatic seismic support design provided by this expert system achieves in the reduction of design man-hours, improvement of design quality, verification of design result, optimization of support locations and prevention of input duplication. In the development of this system, we had to derive the design logic from expert design engineers and this could not be simply expressed descriptively. Also we had to make programs for different kinds of design knowledge. For these reasons we adopted the object oriented programming paradigm (Smalltalk-80) which is suitable for combining programs and carrying out the design work

  18. Texas flexible pavements and overlays : year 1 report, test sections, data collection, analyses, and data storage system.

    Science.gov (United States)

    2012-06-01

    This five-year project was initiated to collect materials and pavement performance data on a minimum of 100 highway test sections around the State of Texas, incorporating both flexible pavements and overlays. Besides being used to calibrate and valid...

  19. Work Zone Data Collection Trailer

    Data.gov (United States)

    Federal Laboratory Consortium — The Work Zone Data Collection Trailer was designed and constructed to enhance data collection and analysis capabilities for the "Evaluating Roadway Construction Work...

  20. An automatic hinge system for leg orthoses

    NARCIS (Netherlands)

    Rietman, J. S.; Goudsmit, J.; Meulemans, D.; Halbertsma, J. P. K.; Geertzen, J. H. B.

    2004-01-01

    This paper describes a new automatic hinge system for leg orthoses, which provides knee stability in stance, and allows knee-flexion during swing. Indications for the hinge system are a paresis or paralysis of the quadriceps muscles. Instrumented gait analysis was performed in three patients, fitted

  1. An automatic hinge system for leg orthoses

    NARCIS (Netherlands)

    Rietman, J.S.; Goudsmit, J.; Meulemans, D.; Halbertsma, J.P.K.; Geertzen, J.H.B.

    This paper describes a new, automatic hinge system for leg orthoses, which provides knee stability in stance, and allows knee-flexion during swing. Indications for the hinge system are a paresis or paralysis of the quadriceps muscles. Instrumented gait analysis was performed in three patients,

  2. Automatic Water Sensor Window Opening System

    KAUST Repository

    Percher, Michael

    2013-01-01

    A system can automatically open at least one window of a vehicle when the vehicle is being submerged in water. The system can include a water collector and a water sensor, and when the water sensor detects water in the water collector, at least one window of the vehicle opens.

  3. Automatic Water Sensor Window Opening System

    KAUST Repository

    Percher, Michael

    2013-12-05

    A system can automatically open at least one window of a vehicle when the vehicle is being submerged in water. The system can include a water collector and a water sensor, and when the water sensor detects water in the water collector, at least one window of the vehicle opens.

  4. Usage of aids monitoring in automatic braking systems of modern cars

    OpenAIRE

    Dembitskyi V.; Mazylyuk P.; Sitovskyi O.

    2016-01-01

    Increased safety can be carried out at the expense the installation on vehicles of automatic braking systems, that monitor the traffic situation and the actions of the driver. In this paper considered the advantages and disadvantages of automatic braking systems, were analyzed modern tracking tools that are used in automatic braking systems. Based on the statistical data on accidents, are set the main dangers, that the automatic braking system will be reduced. In order to ensure the acc...

  5. Temperature Profile Data Collected by Participating Ships in NOAA's Shipboard Environmental Data Acquisition System Program from 17 June 2000 to 23 February 2001 (NODC Accession 0000417)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — XBT and other data were collected from the COLUMBUS COROMANDEL and other platforms participating in NOAA's Shipboard Environmental Data Acquisition System (SEAS)...

  6. 49 CFR Appendix H to Part 40 - DOT Drug and Alcohol Testing Management Information System (MIS) Data Collection Form

    Science.gov (United States)

    2010-10-01

    ..., App. H Appendix H to Part 40—DOT Drug and Alcohol Testing Management Information System (MIS) Data... 49 Transportation 1 2010-10-01 2010-10-01 false DOT Drug and Alcohol Testing Management Information System (MIS) Data Collection Form H Appendix H to Part 40 Transportation Office of the Secretary...

  7. Decision analysis for a data collection system of patient-controlled analgesia with a multi-attribute utility model.

    Science.gov (United States)

    Lee, I-Jung; Huang, Shih-Yu; Tsou, Mei-Yung; Chan, Kwok-Hon; Chang, Kuang-Yi

    2010-10-01

    Data collection systems are very important for the practice of patient-controlled analgesia (PCA). This study aimed to evaluate 3 PCA data collection systems and selected the most favorable system with the aid of multiattribute utility (MAU) theory. We developed a questionnaire with 10 items to evaluate the PCA data collection system and 1 item for overall satisfaction based on MAU theory. Three systems were compared in the questionnaire, including a paper record, optic card reader and personal digital assistant (PDA). A pilot study demonstrated a good internal and test-retest reliability of the questionnaire. A weighted utility score combining the relative importance of individual items assigned by each participant and their responses to each question was calculated for each system. Sensitivity analyses with distinct weighting protocols were conducted to evaluate the stability of the final results. Thirty potential users of a PCA data collection system were recruited in the study. The item "easy to use" had the highest median rank and received the heaviest mean weight among all items. MAU analysis showed that the PDA system had a higher utility score than that in the other 2 systems. Sensitivity analyses revealed that both inverse and reciprocal weighting processes favored the PDA system. High correlations between overall satisfaction and MAU scores from miscellaneous weighting protocols suggested a good predictive validity of our MAU-based questionnaire. The PDA system was selected as the most favorable PCA data collection system by the MAU analysis. The item "easy to use" was the most important attribute of the PCA data collection system. MAU theory can evaluate alternatives by taking into account individual preferences of stakeholders and aid in better decision-making. Copyright © 2010 Elsevier. Published by Elsevier B.V. All rights reserved.

  8. Overcoming Language and Literacy Barriers: Using Student Response System Technology to Collect Quality Program Evaluation Data from Immigrant Participants

    Science.gov (United States)

    Walker, Susan K.; Mao, Dung

    2016-01-01

    Student response system technology was employed for parenting education program evaluation data collection with Karen adults. The technology, with translation and use of an interpreter, provided an efficient and secure method that respected oral language and collective learning preferences and accommodated literacy needs. The method was popular…

  9. From Automatic to Adaptive Data Acquisition

    DEFF Research Database (Denmark)

    Chang, Marcus

    2009-01-01

    the exibility of sensornets and reduce the complexity for the domain scientist, we developed an AI-based controller to act as a proxy between the scientist and sensornet. This controller is driven by the scientist's requirements to the collected data, and uses adaptive sampling in order to reach these goals....

  10. Research on automatic control system of greenhouse

    Science.gov (United States)

    Liu, Yi; Qi, Guoyang; Li, Zeyu; Wu, Qiannan; Meng, Yupeng

    2017-03-01

    This paper introduces a kind of automatic control system of single-chip microcomputer and a temperature and humidity sensor based on the greenhouse, describes the system's hardware structure, working principle and process, and a large number of experiments on the effect of the control system, the results show that the system can ideally control temperature and room temperature and humidity, can be used in indoor breeding and planting, and has the versatility and portability.

  11. Study on intermediate frequency power supply automatic monitor system

    International Nuclear Information System (INIS)

    Wang Yuntong; Xu Bin

    2007-06-01

    A new design project of the automatic monitor system for the intermediate frequency power supply system by using the communication server is put for- ward and the realizing principle method and the key technique are clarified in detail. This system made use of the conversion function with the series communication server's control, realized the data collecting function by the double machine backup and redundancy. The new network system adopted the photoelectric-insulated-communication connect device and the diagnosis technique, increased the anti-interference ability, the communication adopted the technique by the alarm information sending out in first and circularly repeating, the slowly speed is overcame in the original monitor network system, and strengthened the celerity of the monitor system and the reliability of the alarm report. After the new monitor system running, the result shows that the functions is more perfect than the original monitor system, the usage is more convenient, have the higher and dependable stability, the report of alarm is more quickly, and is convenient for the analysis after the trouble, at the same time, the system still have the strong ability and value to expand. (authors)

  12. System for automatic crate recognition

    Directory of Open Access Journals (Sweden)

    Radovan Kukla

    2012-01-01

    Full Text Available This contribution describes usage of computer vision and artificial intelligence methods for application. The method solves abuse of reverse vending machine. This topic has been solved as innovation voucher for the South Moravian Region. It was developed by Mendel university in Brno (Department of informatics – Faculty of Business and Economics and Department of Agricultural, Food and Environmental Engineering – Faculty of Agronomy together with the Czech subsidiary of Tomra. The project is focused on a possibility of integration industrial cameras and computers to process recognition of crates in the verse vending machine. The aim was the effective security system that will be able to save hundreds-thousands financial loss. As suitable development and runtime platform there was chosen product ControlWeb and VisionLab developed by Moravian Instruments Inc.

  13. 75 FR 27001 - Comment Request for Information Collection for the SCSEP Data Collection System, OMB Control No...

    Science.gov (United States)

    2010-05-13

    ... understood, and the impact of collection requirements on respondents can be properly assessed. Currently, the... 2002 (Pub. L. 07-288); changes in overall burden for some forms based on actual usage statistics; and the requirement to publish changes to the Internet-based SCSEP Performance and Results QPR (SPARQ...

  14. Automatic collection of the rare-earths with post chromatography column detection

    International Nuclear Information System (INIS)

    David, P.; Metzger, G.; Repellin, M.

    1987-01-01

    The complete separation of rare-earths (in the aim of radio-isotopes measurement) requires High Performance Liquid Chromatography with ternary elution gradient. To automatize their collection with satisfying conditions, we have realized a non polluting, reliable and easy to operate detection method. This one is based on a derivation colorimetric system with arsenazo I (3 -(2 arsophenylazo 4.5) - dihydroxy - 2.7 naphtalene disulfonic acid)

  15. Automatic oscillator frequency control system

    Science.gov (United States)

    Smith, S. F. (Inventor)

    1985-01-01

    A frequency control system makes an initial correction of the frequency of its own timing circuit after comparison against a frequency of known accuracy and then sequentially checks and corrects the frequencies of several voltage controlled local oscillator circuits. The timing circuit initiates the machine cycles of a central processing unit which applies a frequency index to an input register in a modulo-sum frequency divider stage and enables a multiplexer to clock an accumulator register in the divider stage with a cyclical signal derived from the oscillator circuit being checked. Upon expiration of the interval, the processing unit compares the remainder held as the contents of the accumulator against a stored zero error constant and applies an appropriate correction word to a correction stage to shift the frequency of the oscillator being checked. A signal from the accumulator register may be used to drive a phase plane ROM and, with periodic shifts in the applied frequency index, to provide frequency shift keying of the resultant output signal. Interposition of a phase adder between the accumulator register and phase plane ROM permits phase shift keying of the output signal by periodic variation in the value of a phase index applied to one input of the phase adder.

  16. FY 1983 annual report on the research and development of automatic sewing systems. Total systems; 1983 nendo jido hosei system no kenkyu kaihatsu seika hokokusho. Total system

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1984-03-01

    The automatic sewing system technique research association has been commissioned by the Agency of Industrial Science and Technology for (research and development of automatic sewing systems). This program covers R and D of the elementary techniques for total systems and sewing preparation/processing, sewing/assembling, cloth handling, and system management/control. This report describes the results of the R and D efforts for the total systems. The basic data are collected for a total of 5 representative types of apparels. The technical problems involved in the current and automatic sewing systems are clarified for each apparel type, to collect the data for studying the test plant plans. The automatic sewing system for each apparel type is illustrated by images, for expressing the system concepts. The concept for the basic designs of the total system is drawn for FMS. For achieving the objective of reducing time for producing unit quantity of diversified types of clothes in small quantities at least by 50% from the current level, the study is focused on how the works are (flown) in the system, for which configuration of each device is studied. The basic specifications are drawn, viewed from development of the total system. (NEDO)

  17. Automatic Road Sign Inventory Using Mobile Mapping Systems

    Science.gov (United States)

    Soilán, M.; Riveiro, B.; Martínez-Sánchez, J.; Arias, P.

    2016-06-01

    The periodic inspection of certain infrastructure features plays a key role for road network safety and preservation, and for developing optimal maintenance planning that minimize the life-cycle cost of the inspected features. Mobile Mapping Systems (MMS) use laser scanner technology in order to collect dense and precise three-dimensional point clouds that gather both geometric and radiometric information of the road network. Furthermore, time-stamped RGB imagery that is synchronized with the MMS trajectory is also available. In this paper a methodology for the automatic detection and classification of road signs from point cloud and imagery data provided by a LYNX Mobile Mapper System is presented. First, road signs are detected in the point cloud. Subsequently, the inventory is enriched with geometrical and contextual data such as orientation or distance to the trajectory. Finally, semantic content is given to the detected road signs. As point cloud resolution is insufficient, RGB imagery is used projecting the 3D points in the corresponding images and analysing the RGB data within the bounding box defined by the projected points. The methodology was tested in urban and road environments in Spain, obtaining global recall results greater than 95%, and F-score greater than 90%. In this way, inventory data is obtained in a fast, reliable manner, and it can be applied to improve the maintenance planning of the road network, or to feed a Spatial Information System (SIS), thus, road sign information can be available to be used in a Smart City context.

  18. The WAIS Melt Monitor: An automated ice core melting system for meltwater sample handling and the collection of high resolution microparticle size distribution data

    Science.gov (United States)

    Breton, D. J.; Koffman, B. G.; Kreutz, K. J.; Hamilton, G. S.

    2010-12-01

    Paleoclimate data are often extracted from ice cores by careful geochemical analysis of meltwater samples. The analysis of the microparticles found in ice cores can also yield unique clues about atmospheric dust loading and transport, dust provenance and past environmental conditions. Determination of microparticle concentration, size distribution and chemical makeup as a function of depth is especially difficult because the particle size measurement either consumes or contaminates the meltwater, preventing further geochemical analysis. Here we describe a microcontroller-based ice core melting system which allows the collection of separate microparticle and chemistry samples from the same depth intervals in the ice core, while logging and accurately depth-tagging real-time electrical conductivity and particle size distribution data. This system was designed specifically to support microparticle analysis of the WAIS Divide WDC06A deep ice core, but many of the subsystems are applicable to more general ice core melting operations. Major system components include: a rotary encoder to measure ice core melt displacement with 0.1 millimeter accuracy, a meltwater tracking system to assign core depths to conductivity, particle and sample vial data, an optical debubbler level control system to protect the Abakus laser particle counter from damage due to air bubbles, a Rabbit 3700 microcontroller which communicates with a host PC, collects encoder and optical sensor data and autonomously operates Gilson peristaltic pumps and fraction collectors to provide automatic sample handling, melt monitor control software operating on a standard PC allowing the user to control and view the status of the system, data logging software operating on the same PC to collect data from the melting, electrical conductivity and microparticle measurement systems. Because microparticle samples can easily be contaminated, we use optical air bubble sensors and high resolution ice core density

  19. Recent developments in the Los Alamos National Laboratory Plutonium Facility Waste Tracking System-automated data collection pilot project

    International Nuclear Information System (INIS)

    Martinez, B.; Montoya, A.; Klein, W.

    1999-01-01

    The waste management and environmental compliance group (NMT-7) at the Los Alamos National Laboratory has initiated a pilot project for demonstrating the feasibility and utility of automated data collection as a solution for tracking waste containers at the Los Alamos National Laboratory Plutonium Facility. This project, the Los Alamos Waste Tracking System (LAWTS), tracks waste containers during their lifecycle at the facility. LAWTS is a two-tiered system consisting of a server/workstation database and reporting engine and a hand-held data terminal-based client program for collecting data directly from tracked containers. New containers may be added to the system from either the client unit or from the server database. Once containers are in the system, they can be tracked through one of three primary transactions: Move, Inventory, and Shipment. Because LAWTS is a pilot project, it also serves as a learning experience for all parties involved. This paper will discuss many of the lessons learned in implementing a data collection system in the restricted environment. Specifically, the authors will discuss issues related to working with the PPT 4640 terminal system as the data collection unit. They will discuss problems with form factor (size, usability, etc.) as well as technical problems with wireless radio frequency functions. They will also discuss complications that arose from outdoor use of the terminal (barcode scanning failures, screen readability problems). The paper will conclude with a series of recommendations for proceeding with LAWTS based on experience to date

  20. A data management program for the Electra 800 automatic analyser.

    Science.gov (United States)

    Cambus, J P; Nguyen, F; de Graeve, J; Aragon, B; Valdiguie, P

    1994-10-01

    The Electra 800 automatic coagulation analyser rapidly performs most chronometric coagulation tests with high precision. To facilitate data handling, software, adaptable to any PC running under MS-DOS, was written to manage the analyser. Data are automatically collected via the RS232 interface or can be manually input. The software can handle 64 different analyses, all entirely 'user defined'. An 'electronic worksheet' presents the results in pages of ten patients. This enables the operator to assess the data and to perform verifications or complementary tests if necessary. All results outside a predetermined range can be flagged and results can be deleted, modified or added. A patient's previous files can be recalled as the data are archived at the end of the day. A 120 Mb disk can store approximately 130,000 patient files. A daily archive function can print the day's work in alphabetical order. A communication protocol allows connection to a mainframe computer. This program and the user's manual are available on request, free of charge, from the authors.

  1. Learning Diagnostic Diagrams in Transport-Based Data-Collection Systems

    DEFF Research Database (Denmark)

    Tran, Vu The; Eklund, Peter; Cook, Chris

    2014-01-01

    Insights about service improvement in a transit network can be gained by studying transit service reliability. In this paper, a general procedure for constructing a transit service reliability diagnostic (Tsrd) diagram based on a Bayesian network is proposed to automatically build a behavioural...

  2. Two Systems for Automatic Music Genre Recognition

    DEFF Research Database (Denmark)

    Sturm, Bob L.

    2012-01-01

    We re-implement and test two state-of-the-art systems for automatic music genre classification; but unlike past works in this area, we look closer than ever before at their behavior. First, we look at specific instances where each system consistently applies the same wrong label across multiple...... trials of cross-validation. Second, we test the robustness of each system to spectral equalization. Finally, we test how well human subjects recognize the genres of music excerpts composed by each system to be highly genre representative. Our results suggest that neither high-performing system has...... a capacity to recognize music genre....

  3. Towards Automatic Capturing of Manual Data Processing Provenance

    NARCIS (Netherlands)

    Wombacher, Andreas; Huq, M.R.

    2011-01-01

    Often data processing is not implemented by a work ow system or an integration application but is performed manually by humans along the lines of a more or less specified procedure. Collecting provenance information during manual data processing can not be automated. Further, manual collection of

  4. Development of an automatic human duress detection system

    International Nuclear Information System (INIS)

    Greene, E.R.; Davis, J.G.; Tuttle, W.C.

    1979-01-01

    A method for automatically detecting duress in security personnel utilizes real-time assessment of physiological data (heart rate) to evaluate psychological stress. Using body-worn tape recorders, field data have been collected on 22 Albuquerque police officers (20 male, 2 female) to determine actual heart rate responses in both routine and life-threatening situations. Off-line computer analysis has been applied to the data to determine the speed and reliability with which an alarm could be triggered. Alarm algorithms relating field responses to laboratory collected baseline responses have been developed

  5. Water quality data collected by the the National Estuarine Research Reserve System's System-wide Monitoring Program (NERRS SWMP), 1996 - 1998 (NODC Accession 0000789)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The National Estuarine Research Reserve System's System-wide Monitoring Program (NERRS SWMP) collected water quality data in 22 reserves in the United States and...

  6. Water quality, meteorological, and nutrient data collected by the the National Estuarine Research Reserve System's System-wide Monitoring Program (NERRS SWMP), 1994 - 2005 (NODC Accession 0019215)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The National Estuarine Research Reserve System's System-wide Monitoring Program (NERRS SWMP) collected water quality, meteorological, and nutrient data in 25...

  7. A versatile system for the rapid collection, handling and graphics analysis of multidimensional data

    International Nuclear Information System (INIS)

    O'Brien, P.M.; Moloney, G.; O'Oconnor, A.; Legge, G.J.F.

    1991-01-01

    The paper discusses the performances of a versatile computerized system developed at the Microanalytical Research Centre of the Melbourne University, for handling multiparameter data that may arise from a variety of experiments - nuclear, accelerator mass spectrometry, microprobe elemental analysis or 3-D microtomography. Some of the most demanding requirements arise in the application of microprobes to quantitative elemental mapping and to microtomography. A system to handle data from such experiments had been under continuous development. It has been reprogramed to run on a DG DS7540 workstation. The whole system of software has been rewritten, greatly expanded and made much more powerful and faster, by use of modern computer technology - a VME bus computer with a real-time operating system and a RISC workstation running UNIX and the X-window environment

  8. Travel time data collection handbook

    Science.gov (United States)

    1998-03-01

    This Travel Time Data Collection Handbook provides guidance to transportation : professionals and practitioners for the collection, reduction, and presentation : of travel time data. The handbook should be a useful reference for designing : travel ti...

  9. A Risk Assessment System with Automatic Extraction of Event Types

    Science.gov (United States)

    Capet, Philippe; Delavallade, Thomas; Nakamura, Takuya; Sandor, Agnes; Tarsitano, Cedric; Voyatzi, Stavroula

    In this article we describe the joint effort of experts in linguistics, information extraction and risk assessment to integrate EventSpotter, an automatic event extraction engine, into ADAC, an automated early warning system. By detecting as early as possible weak signals of emerging risks ADAC provides a dynamic synthetic picture of situations involving risk. The ADAC system calculates risk on the basis of fuzzy logic rules operated on a template graph whose leaves are event types. EventSpotter is based on a general purpose natural language dependency parser, XIP, enhanced with domain-specific lexical resources (Lexicon-Grammar). Its role is to automatically feed the leaves with input data.

  10. Real time psychrometric data collection

    International Nuclear Information System (INIS)

    McDaniel, K.H.

    1996-01-01

    Eight Mine Weather Stations (MWS) installed at the Waste Isolation Pilot Plant (WIPP) to monitor the underground ventilation system are helping to simulate real-time ventilation scenarios. Seasonal weather extremes can result in variations of Natural Ventilation Pressure (NVP) which can significantly effect the ventilation system. The eight MWS(s) (which previously collected and stored temperature, barometric pressure and relative humidity data for subsequent NVP calculations) were upgraded to provide continuous real-time data to the site wide Central monitoring System. This data can now be utilized by the ventilation engineer to create realtime ventilation simulations and trends which assist in the prediction and mitigation of NVP and psychrometric related events

  11. Taxing the cloud: introducing a new taxation system on data collection?

    Directory of Open Access Journals (Sweden)

    Primavera De Filippi

    2013-05-01

    Full Text Available Cloud computing services are increasingly hosted on international servers and distributed amongst multiple data centres. Given their global scope, it is often easier for large multinational corporations to effectively circumvent old taxation schemes designed around the concept of territorial jurisdiction and geographical settings. In view of obtaining tax revenues from these online operators whose business is partially carried out in France, the French government recently issued a report emphasising the need for new taxation rules that would better comply with the way value is generated in the digital economy: at the international level, it is suggested that taxation should be calculated according to the place of interaction with end-users; at the national level, the report suggests to introduce a transitory tax on data collection in order to promote innovation and encourage good online practices.

  12. 49 CFR 236.825 - System, automatic train control.

    Science.gov (United States)

    2010-10-01

    ... 49 Transportation 4 2010-10-01 2010-10-01 false System, automatic train control. 236.825 Section..., INSPECTION, MAINTENANCE, AND REPAIR OF SIGNAL AND TRAIN CONTROL SYSTEMS, DEVICES, AND APPLIANCES Definitions § 236.825 System, automatic train control. A system so arranged that its operation will automatically...

  13. DMET-analyzer: automatic analysis of Affymetrix DMET data.

    Science.gov (United States)

    Guzzi, Pietro Hiram; Agapito, Giuseppe; Di Martino, Maria Teresa; Arbitrio, Mariamena; Tassone, Pierfrancesco; Tagliaferri, Pierosandro; Cannataro, Mario

    2012-10-05

    Clinical Bioinformatics is currently growing and is based on the integration of clinical and omics data aiming at the development of personalized medicine. Thus the introduction of novel technologies able to investigate the relationship among clinical states and biological machineries may help the development of this field. For instance the Affymetrix DMET platform (drug metabolism enzymes and transporters) is able to study the relationship among the variation of the genome of patients and drug metabolism, detecting SNPs (Single Nucleotide Polymorphism) on genes related to drug metabolism. This may allow for instance to find genetic variants in patients which present different drug responses, in pharmacogenomics and clinical studies. Despite this, there is currently a lack in the development of open-source algorithms and tools for the analysis of DMET data. Existing software tools for DMET data generally allow only the preprocessing of binary data (e.g. the DMET-Console provided by Affymetrix) and simple data analysis operations, but do not allow to test the association of the presence of SNPs with the response to drugs. We developed DMET-Analyzer a tool for the automatic association analysis among the variation of the patient genomes and the clinical conditions of patients, i.e. the different response to drugs. The proposed system allows: (i) to automatize the workflow of analysis of DMET-SNP data avoiding the use of multiple tools; (ii) the automatic annotation of DMET-SNP data and the search in existing databases of SNPs (e.g. dbSNP), (iii) the association of SNP with pathway through the search in PharmaGKB, a major knowledge base for pharmacogenomic studies. DMET-Analyzer has a simple graphical user interface that allows users (doctors/biologists) to upload and analyse DMET files produced by Affymetrix DMET-Console in an interactive way. The effectiveness and easy use of DMET Analyzer is demonstrated through different case studies regarding the analysis of

  14. DMET-Analyzer: automatic analysis of Affymetrix DMET Data

    Directory of Open Access Journals (Sweden)

    Guzzi Pietro

    2012-10-01

    Full Text Available Abstract Background Clinical Bioinformatics is currently growing and is based on the integration of clinical and omics data aiming at the development of personalized medicine. Thus the introduction of novel technologies able to investigate the relationship among clinical states and biological machineries may help the development of this field. For instance the Affymetrix DMET platform (drug metabolism enzymes and transporters is able to study the relationship among the variation of the genome of patients and drug metabolism, detecting SNPs (Single Nucleotide Polymorphism on genes related to drug metabolism. This may allow for instance to find genetic variants in patients which present different drug responses, in pharmacogenomics and clinical studies. Despite this, there is currently a lack in the development of open-source algorithms and tools for the analysis of DMET data. Existing software tools for DMET data generally allow only the preprocessing of binary data (e.g. the DMET-Console provided by Affymetrix and simple data analysis operations, but do not allow to test the association of the presence of SNPs with the response to drugs. Results We developed DMET-Analyzer a tool for the automatic association analysis among the variation of the patient genomes and the clinical conditions of patients, i.e. the different response to drugs. The proposed system allows: (i to automatize the workflow of analysis of DMET-SNP data avoiding the use of multiple tools; (ii the automatic annotation of DMET-SNP data and the search in existing databases of SNPs (e.g. dbSNP, (iii the association of SNP with pathway through the search in PharmaGKB, a major knowledge base for pharmacogenomic studies. DMET-Analyzer has a simple graphical user interface that allows users (doctors/biologists to upload and analyse DMET files produced by Affymetrix DMET-Console in an interactive way. The effectiveness and easy use of DMET Analyzer is demonstrated through different

  15. Gamma-ray spectrometry data collection and reduction by simple computing systems

    International Nuclear Information System (INIS)

    Op de Beeck, J.

    1975-01-01

    The review summarizes the present state of the involvement of relatively small computing devices in the collection and processing of gamma-ray spectrum data. An economic and utilitarian point of view has been chosen with regard to data collection in order to arrive at practically valuable conclusions in terms of feasibility of possible configurations with respect to their eventual application. A unified point of view has been adopted with regard to data processing by developing an information theoretical approach on a more or less intuitive level in an attempt to remove the largest part of the virtual disparity between the several processing methods described in the literature. A synoptical introduction to the most important mathematical methods has been incorporated, together with a detailed theoretical description of the concept gamma-ray spectrum. In accordance with modern requirements, the discussions are mainly oriented towards high-resolution semiconductor detector-type spectra. The critical evaluation of the processing methods reviewed is done with respect to a set of predefined criteria. Smoothing, peak detection, peak intensity determination, overlapping peak resolving and detection and upper limits are discussed in great detail. A preferred spectrum analysis method combining powerful data reduction properties with extreme simplicity and speed of operation is suggested. The general discussion is heavily oriented towards activation analysis application, but other disciplines making use of gamma-ray spectrometry will find the material presented equally useful. Final conclusions are given pointing to future developments and shifting their centre of gravity towards improving the quality of the measurements rather than expanding the use of tedious and sophisticated mathematical techniques requiring the limits of available computational power. (author)

  16. Wireless data collection system for travel time estimation and traffic performance evaluation.

    Science.gov (United States)

    2010-09-01

    Having accurate and continually updated travel time and other performance data for the road and highway system has many benefits. From the perspective of the road users, having real-time updates on travel times will permit better travel and route pla...

  17. Method of software development for tasks of automatic control systems for simulation and designing on the base of the technological systems design data

    International Nuclear Information System (INIS)

    Ajzatulin, A.I.

    2007-01-01

    One studies the factors affecting the designing of the full-scale simulation facilities, the design data base simulation and the application of digital computerized process control systems. Paper describes problems dealing with the errors in the process system design data and the algorithm simulation methodological problems. On the basis of the records of the efforts to design the full-scale simulation facilities of the Tienvan NPP and of the Kudankulam NPP one brings to the notice a procedure to elaborate new tools to simulate and to elaborate algorithms for the computerized process control systems based on the process system design data. Paper lists the basic components of the program system under elaboration to ensure simulation and designing and describes their functions. The introduction result is briefly described [ru

  18. Automatic Anthropometric System Development Using Machine Learning

    Directory of Open Access Journals (Sweden)

    Long The Nguyen

    2016-08-01

    Full Text Available The contactless automatic anthropometric system is proposed for the reconstruction of the 3D-model of the human body using the conventional smartphone. Our approach involves three main steps. The first step is the extraction of 12 anthropological features. Then we determine the most important features. Finally, we employ these features to build the 3D model of the human body and classify them according to gender and the commonly used sizes. 

  19. PLC Based Automatic Multistoried Car Parking System

    OpenAIRE

    Swanand S .Vaze; Rohan S. Mithari

    2014-01-01

    This project work presents the study and design of PLC based Automatic Multistoried Car Parking System. Multistoried car parking is an arrangement which is used to park a large number of vehicles in least possible place. For making this arrangement in a real plan very high technological instruments are required. In this project a prototype of such a model is made. This prototype model is made for accommodating twelve cars at a time. Availability of the space for parking is detecte...

  20. The Collection of Event Data and its Relevance to the Optimisation of Decay Heat Rejection Systems

    International Nuclear Information System (INIS)

    Roughley, R.; Jones, N.

    1975-01-01

    The precision with which the reliability of DHR (Decay Heat Rejection) systems for nuclear reactors can be predicted depends not only upon model representation but also on the accuracy of the data used. In the preliminary design stages when models are being used to arrive at major engineering decisions in relation to plant configuration, the best the designer can do is use the data available at the time. With the present state of the art it is acknowledged that some degree of judgement will have to be exercised particularly for plant involving sodium technology where a large amount of operational experience has not yet been generated. This paper reviews the current efforts being deployed in the acquisition of field data relevant to DHR systems so that improvements in reliability predictions may be realised

  1. Automatic analysis of attack data from distributed honeypot network

    Science.gov (United States)

    Safarik, Jakub; Voznak, MIroslav; Rezac, Filip; Partila, Pavol; Tomala, Karel

    2013-05-01

    There are many ways of getting real data about malicious activity in a network. One of them relies on masquerading monitoring servers as a production one. These servers are called honeypots and data about attacks on them brings us valuable information about actual attacks and techniques used by hackers. The article describes distributed topology of honeypots, which was developed with a strong orientation on monitoring of IP telephony traffic. IP telephony servers can be easily exposed to various types of attacks, and without protection, this situation can lead to loss of money and other unpleasant consequences. Using a distributed topology with honeypots placed in different geological locations and networks provides more valuable and independent results. With automatic system of gathering information from all honeypots, it is possible to work with all information on one centralized point. Communication between honeypots and centralized data store use secure SSH tunnels and server communicates only with authorized honeypots. The centralized server also automatically analyses data from each honeypot. Results of this analysis and also other statistical data about malicious activity are simply accessible through a built-in web server. All statistical and analysis reports serve as information basis for an algorithm which classifies different types of used VoIP attacks. The web interface then brings a tool for quick comparison and evaluation of actual attacks in all monitored networks. The article describes both, the honeypots nodes in distributed architecture, which monitor suspicious activity, and also methods and algorithms used on the server side for analysis of gathered data.

  2. Computerized ECT data analysis system

    International Nuclear Information System (INIS)

    Miyake, Y.; Fukui, S.; Iwahashi, Y.; Matsumoto, M.; Koyama, K.

    1988-01-01

    For the analytical method of the eddy current testing (ECT) of steam generator tubes in nuclear power plants, the authors have developed the computerized ECT data analysis system using a large-scale computer with a high-resolution color graphic display. This system can store acquired ECT data up to 15 steam generators, and ECT data can be analyzed immediately on the monitor in dialogue communication with a computer. Analyzed results of ECT data are stored and registered in the data base. This system enables an analyst to perform sorting and collecting of data under various conditions and obtain the results automatically, and also to make a plan of tube repair works. This system has completed the test run, and has been used for data analysis at the annual inspection of domestic plants. This paper describes an outline, features and examples of the computerized eddy current data analysis system for steam generator tubes in PWR nuclear power plants

  3. Design of USB/RS485 converter and its application in slow control data collection system of high energy physics

    International Nuclear Information System (INIS)

    Chen Xihui; Xie Song; Gao Cuishan; Xie Xiaoxi; Nie Zhendong; Zhang Yinhong; Gao Lu

    2005-01-01

    Most of traditional data collection systems are based on RS232/485 converter. Such a system can only realize the point-to-point connection and its branches cannot work independently. So it brings on lots of inconvenience in debugging, installing and maintenance. On the other hand, the widely used Universal Serial Bus (USB) has many advantages such as hot-plugging, easy to extend, convenient in installing and occupying less system resources. So if USB could be used in data collection system, it would bring on lots of convenience to the system and its branches could be work independently. The design of USB-485 converter and its application are introduced in this paper. (authors)

  4. Intelligent Management System of Power Network Information Collection Under Big Data Storage

    Directory of Open Access Journals (Sweden)

    Qin Yingying

    2017-01-01

    Full Text Available With the development of economy and society, big data storage in enterprise management has become a problem that can’t be ignored. How to manage and optimize the allocation of tasks better is an important factor in the sustainable development of an enterprise. Now the enterprise information intelligent management has become a hot spot of management mode and concept in the information age. It presents information to the business managers in a more efficient, lower cost, and global form. The system uses the SG-UAP development tools, which is based on Eclipse development environment, and suits for Windows operating system, with Oracle as database development platform, Tomcat network information service for application server. The system uses SOA service-oriented architecture, provides RESTful style service, and HTTP(S as the communication protocol, and JSON as the data format. The system is divided into two parts, the front-end and the backs-end, achieved functions like user login, registration, password retrieving, enterprise internal personnel information management and internal data display and other functions.

  5. Automatic code generation for distributed robotic systems

    International Nuclear Information System (INIS)

    Jones, J.P.

    1993-01-01

    Hetero Helix is a software environment which supports relatively large robotic system development projects. The environment supports a heterogeneous set of message-passing LAN-connected common-bus multiprocessors, but the programming model seen by software developers is a simple shared memory. The conceptual simplicity of shared memory makes it an extremely attractive programming model, especially in large projects where coordinating a large number of people can itself become a significant source of complexity. We present results from three system development efforts conducted at Oak Ridge National Laboratory over the past several years. Each of these efforts used automatic software generation to create 10 to 20 percent of the system

  6. A Context Dependent Automatic Target Recognition System

    Science.gov (United States)

    Kim, J. H.; Payton, D. W.; Olin, K. E.; Tseng, D. Y.

    1984-06-01

    This paper describes a new approach to automatic target recognizer (ATR) development utilizing artificial intelligent techniques. The ATR system exploits contextual information in its detection and classification processes to provide a high degree of robustness and adaptability. In the system, knowledge about domain objects and their contextual relationships is encoded in frames, separating it from low level image processing algorithms. This knowledge-based system demonstrates an improvement over the conventional statistical approach through the exploitation of diverse forms of knowledge in its decision-making process.

  7. Automation of chromosomes analysis. Automatic system for image processing

    International Nuclear Information System (INIS)

    Le Go, R.; Cosnac, B. de; Spiwack, A.

    1975-01-01

    The A.S.T.I. is an automatic system relating to the fast conversational processing of all kinds of images (cells, chromosomes) converted to a numerical data set (120000 points, 16 grey levels stored in a MOS memory) through a fast D.O. analyzer. The system performs automatically the isolation of any individual image, the area and weighted area of which are computed. These results are directly displayed on the command panel and can be transferred to a mini-computer for further computations. A bright spot allows parts of an image to be picked out and the results to be displayed. This study is particularly directed towards automatic karyo-typing [fr

  8. An Automatic Indirect Immunofluorescence Cell Segmentation System

    Directory of Open Access Journals (Sweden)

    Yung-Kuan Chan

    2014-01-01

    Full Text Available Indirect immunofluorescence (IIF with HEp-2 cells has been used for the detection of antinuclear autoantibodies (ANA in systemic autoimmune diseases. The ANA testing allows us to scan a broad range of autoantibody entities and to describe them by distinct fluorescence patterns. Automatic inspection for fluorescence patterns in an IIF image can assist physicians, without relevant experience, in making correct diagnosis. How to segment the cells from an IIF image is essential in developing an automatic inspection system for ANA testing. This paper focuses on the cell detection and segmentation; an efficient method is proposed for automatically detecting the cells with fluorescence pattern in an IIF image. Cell culture is a process in which cells grow under control. Cell counting technology plays an important role in measuring the cell density in a culture tank. Moreover, assessing medium suitability, determining population doubling times, and monitoring cell growth in cultures all require a means of quantifying cell population. The proposed method also can be used to count the cells from an image taken under a fluorescence microscope.

  9. Automatic-Control System for Safer Brazing

    Science.gov (United States)

    Stein, J. A.; Vanasse, M. A.

    1986-01-01

    Automatic-control system for radio-frequency (RF) induction brazing of metal tubing reduces probability of operator errors, increases safety, and ensures high-quality brazed joints. Unit combines functions of gas control and electric-power control. Minimizes unnecessary flow of argon gas into work area and prevents electrical shocks from RF terminals. Controller will not allow power to flow from RF generator to brazing head unless work has been firmly attached to head and has actuated micro-switch. Potential shock hazard eliminated. Flow of argon for purging and cooling must be turned on and adjusted before brazing power applied. Provision ensures power not applied prematurely, causing damaged work or poor-quality joints. Controller automatically turns off argon flow at conclusion of brazing so potentially suffocating gas does not accumulate in confined areas.

  10. Semi-automatic drawings surveying system

    International Nuclear Information System (INIS)

    Andriamampianina, Lala

    1983-01-01

    A system for the semi-automatic survey of drawings is presented. Its design has been oriented to the reduction of the stored information required for the drawing reproduction. This equipment consists mainly of a plotter driven by a micro-computer, but the pen of the plotter is replaced by a circular photodiode array. Line drawings are first viewed as a concatenation of vectors, with constant angle between the two vectors, and then divided in arcs of circles and line segments. A dynamic analysis of line intersections with the circular sensor permits to identify starting points and end points in a line, for the purpose of automatically following connected lines in drawing. The advantage of the method described is that precision practically depends only on the plotter performance, the sensor resolution being only considered for the thickness of strokes and the distance between two strokes. (author) [fr

  11. 14 CFR 29.1329 - Automatic pilot system.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 1 2010-01-01 2010-01-01 false Automatic pilot system. 29.1329 Section 29... pilot system. (a) Each automatic pilot system must be designed so that the automatic pilot can— (1) Be sufficiently overpowered by one pilot to allow control of the rotorcraft; and (2) Be readily and positively...

  12. 14 CFR 27.1329 - Automatic pilot system.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 1 2010-01-01 2010-01-01 false Automatic pilot system. 27.1329 Section 27... pilot system. (a) Each automatic pilot system must be designed so that the automatic pilot can— (1) Be sufficiently overpowered by one pilot to allow control of the rotorcraft; and (2) Be readily and positively...

  13. Adaptive pseudolinear compensators of dynamic characteristics of automatic control systems

    Science.gov (United States)

    Skorospeshkin, M. V.; Sukhodoev, M. S.; Timoshenko, E. A.; Lenskiy, F. V.

    2016-04-01

    Adaptive pseudolinear gain and phase compensators of dynamic characteristics of automatic control systems are suggested. The automatic control system performance with adaptive compensators has been explored. The efficiency of pseudolinear adaptive compensators in the automatic control systems with time-varying parameters has been demonstrated.

  14. A fully automatic system for acid-base coulometric titrations

    OpenAIRE

    Cladera, A.; Caro, A.; Estela, J. M.; Cerdà, V.

    1990-01-01

    An automatic system for acid-base titrations by electrogeneration of H+ and OH- ions, with potentiometric end-point detection, was developed. The system includes a PC-compatible computer for instrumental control, data acquisition and processing, which allows up to 13 samples to be analysed sequentially with no human intervention. The system performance was tested on the titration of standard solutions, which it carried out with low errors and RSD. It was subsequently applied to the analysis o...

  15. Automatic acoustic and vibration monitoring system for nuclear power plants

    International Nuclear Information System (INIS)

    Tothmatyas, Istvan; Illenyi, Andras; Kiss, Jozsef; Komaromi, Tibor; Nagy, Istvan; Olchvary, Geza

    1990-01-01

    A diagnostic system for nuclear power plant monitoring is described. Acoustic and vibration diagnostics can be applied to monitor various reactor components and auxiliary equipment including primary circuit machinery, leak detection, integrity of reactor vessel, loose parts monitoring. A noise diagnostic system has been developed for the Paks Nuclear Power Plant, to supervise the vibration state of primary circuit machinery. An automatic data acquisition and processing system is described for digitalizing and analysing diagnostic signals. (R.P.) 3 figs

  16. ODM Data Analysis-A tool for the automatic validation, monitoring and generation of generic descriptive statistics of patient data.

    Science.gov (United States)

    Brix, Tobias Johannes; Bruland, Philipp; Sarfraz, Saad; Ernsting, Jan; Neuhaus, Philipp; Storck, Michael; Doods, Justin; Ständer, Sonja; Dugas, Martin

    2018-01-01

    A required step for presenting results of clinical studies is the declaration of participants demographic and baseline characteristics as claimed by the FDAAA 801. The common workflow to accomplish this task is to export the clinical data from the used electronic data capture system and import it into statistical software like SAS software or IBM SPSS. This software requires trained users, who have to implement the analysis individually for each item. These expenditures may become an obstacle for small studies. Objective of this work is to design, implement and evaluate an open source application, called ODM Data Analysis, for the semi-automatic analysis of clinical study data. The system requires clinical data in the CDISC Operational Data Model format. After uploading the file, its syntax and data type conformity of the collected data is validated. The completeness of the study data is determined and basic statistics, including illustrative charts for each item, are generated. Datasets from four clinical studies have been used to evaluate the application's performance and functionality. The system is implemented as an open source web application (available at https://odmanalysis.uni-muenster.de) and also provided as Docker image which enables an easy distribution and installation on local systems. Study data is only stored in the application as long as the calculations are performed which is compliant with data protection endeavors. Analysis times are below half an hour, even for larger studies with over 6000 subjects. Medical experts have ensured the usefulness of this application to grant an overview of their collected study data for monitoring purposes and to generate descriptive statistics without further user interaction. The semi-automatic analysis has its limitations and cannot replace the complex analysis of statisticians, but it can be used as a starting point for their examination and reporting.

  17. Introduction to monitoring dynamic environmental phenomena of the world using satellite data collection systems, 1978

    Science.gov (United States)

    Carter, William Douglas; Paulson, Richard W.

    1979-01-01

    The rapid development of satellite technology, especially in the area of radio transmission and imaging systems, makes it possible to monitor dynamic surface phenomena of the Earth in considerable detail. The monitoring systems that have been developed are compatible with standard monitoring systems such as snow, stream, and rain gages; wind, temperature and humidity measuring instruments; tiltmeters and seismic event counters. Supported by appropriate power, radios and antennae, remote stations can be left unattended for at least 1 year and consistently relay local information via polar orbiting or geostationary satellites. These data, in conjunction with timely Landsat images, can provide a basis for more accurate estimates on snowfall, water runoff, reservoir level changes, flooding, drought effects, and vegetation trends and may be of help in forecasting volcanic eruptions. These types of information are critical for resource inventory and development, especially in developing countries where remote regions are commonly difficult to access. This paper introduces the reader to the systems available, describes their features and limitations, and provides suggestions on how to employ them. An extensive bibliography is provided for those who wish more information.

  18. Investigation of periodically driven systems by x-ray absorption spectroscopy using asynchronous data collection mode

    Science.gov (United States)

    Singh, H.; Donetsky, D.; Liu, J.; Attenkofer, K.; Cheng, B.; Trelewicz, J. R.; Lubomirsky, I.; Stavitski, E.; Frenkel, A. I.

    2018-04-01

    We report the development, testing, and demonstration of a setup for modulation excitation spectroscopy experiments at the Inner Shell Spectroscopy beamline of National Synchrotron Light Source - II. A computer algorithm and dedicated software were developed for asynchronous data processing and analysis. We demonstrate the reconstruction of X-ray absorption spectra for different time points within the modulation pulse using a model system. This setup and the software are intended for a broad range of functional materials which exhibit structural and/or electronic responses to the external stimulation, such as catalysts, energy and battery materials, and electromechanical devices.

  19. Parallel log structured file system collective buffering to achieve a compact representation of scientific and/or dimensional data

    Science.gov (United States)

    Grider, Gary A.; Poole, Stephen W.

    2015-09-01

    Collective buffering and data pattern solutions are provided for storage, retrieval, and/or analysis of data in a collective parallel processing environment. For example, a method can be provided for data storage in a collective parallel processing environment. The method comprises receiving data to be written for a plurality of collective processes within a collective parallel processing environment, extracting a data pattern for the data to be written for the plurality of collective processes, generating a representation describing the data pattern, and saving the data and the representation.

  20. An automatic drawing system for a report radioactive contamination check

    International Nuclear Information System (INIS)

    Saneyoshi, Keiji; Tomita, Satoru; Yoda, Isao

    2002-01-01

    An Automatic drawing system for a report of surface contamination check in a radiation controlled area has been developed. The system can print out the report applied for the format provided by the law from the raw data that is the output from measuring instruments. The task of a worker is only to insert an FD storing the data into a PC and to push a button. The system also yields contamination maps to indicate contamination points clearly. With this system the time to complete the report from the raw data could be decreased from more than two hours to 4 minutes. (author)

  1. Estimating spatial travel times using automatic vehicle identification data

    Science.gov (United States)

    2001-01-01

    Prepared ca. 2001. The paper describes an algorithm that was developed for estimating reliable and accurate average roadway link travel times using Automatic Vehicle Identification (AVI) data. The algorithm presented is unique in two aspects. First, ...

  2. Automatic motion inhibit system for a nuclear power generating system

    International Nuclear Information System (INIS)

    Musick, C.R.; Torres, J.M.

    1977-01-01

    Disclosed is an automatic motion inhibit system for a nuclear power generating system for inhibiting automatic motion of the control elements to reduce reactor power in response to a turbine load reduction. The system generates a final reactor power level setpoint signal which is continuously compared with a reactor power signal. The final reactor power level setpoint is a setpoint within the capacity of the bypass valves to bypass steam which in no event is lower in value than the lower limit of automatic control of the reactor. If the final reactor power level setpoint is greater than the reactor power, an inhibit signal is generated to inhibit automatic control of the reactor. 6 claims, 5 figures

  3. A new approach to configurable primary data collection.

    Science.gov (United States)

    Stanek, J; Babkin, E; Zubov, M

    2016-09-01

    The formats, semantics and operational rules of data processing tasks in genomics (and health in general) are highly divergent and can rapidly change. In such an environment, the problem of consistent transformation and loading of heterogeneous input data to various target repositories becomes a critical success factor. The objective of the project was to design a new conceptual approach to configurable data transformation, de-identification, and submission of health and genomic data sets. Main motivation was to facilitate automated or human-driven data uploading, as well as consolidation of heterogeneous sources in large genomic or health projects. Modern methods of on-demand specialization of generic software components were applied. For specification of input-output data and required data collection activities, we propose a simple data model of flat tables as well as a domain-oriented graphical interface and portable representation of transformations in XML. Using such methods, the prototype of the Configurable Data Collection System (CDCS) was implemented in Java programming language with Swing graphical interfaces. The core logic of transformations was implemented as a library of reusable plugins. The solution is implemented as a software prototype for a configurable service-oriented system for semi-automatic data collection, transformation, sanitization and safe uploading to heterogeneous data repositories-CDCS. To address the dynamic nature of data schemas and data collection processes, the CDCS prototype facilitates interactive, user-driven configuration of the data collection process and extends basic functionality with a wide range of third-party plugins. Notably, our solution also allows for the reduction of manual data entry for data originally missing in the output data sets. First experiments and feedback from domain experts confirm the prototype is flexible, configurable and extensible; runs well on data owner's systems; and is not dependent on

  4. Effectiveness of an automatic manual wheelchair braking system in the prevention of falls.

    Science.gov (United States)

    Martorello, Laura; Swanson, Edward

    2006-01-01

    The purpose of this study was to evaluate the effectiveness of an automatic manual wheelchair braking system in the reduction of falls for patients at high risk of falls while transferring to and from a manual wheelchair. The study design was a normative survey carried out through the use of a written questionnaire sent to 60 skilled nursing facilities to collect data from the medical charts, which identified patients at high risk for falls who used an automatic wheelchair braking system. The facilities participating in the study identified a frequency of falls of high-risk patients while transferring to and from the wheelchair ranging from 2 to 10 per year, with a median fall rate per facility of 4 falls. One year after the installation of the automatic wheelchair braking system, participating facilities demonstrated a reduction of zero to three falls during transfers by high-risk patients, with a median fall rate of zero falls. This represents a statistically significant reduction of 78% in the fall rate of high-risk patients while transferring to and from the wheelchair, t (18) = 6.39, p braking system for manual wheelchairs was installed. The application of the automatic braking system allows clients, families/caregivers, and facility personnel an increased safety factor for the reduction of falls from the wheelchair.

  5. Neural network for automatic analysis of motility data

    DEFF Research Database (Denmark)

    Jakobsen, Erik; Kruse-Andersen, S; Kolberg, Jens Godsk

    1994-01-01

    comparable. However, the neural network recognized pressure peaks clearly generated by muscular activity that had escaped detection by the conventional program. In conclusion, we believe that neurocomputing has potential advantages for automatic analysis of gastrointestinal motility data.......Continuous recording of intraluminal pressures for extended periods of time is currently regarded as a valuable method for detection of esophageal motor abnormalities. A subsequent automatic analysis of the resulting motility data relies on strict mathematical criteria for recognition of pressure...

  6. A method of automatic data processing in radiometric control

    International Nuclear Information System (INIS)

    Adonin, V.M.; Gulyukina, N.A.; Nemirov, Yu.V.; Mogil'nitskij, M.I.

    1980-01-01

    Described is the algorithm for automatic data processing in gamma radiography of products. Rapidity due to application of recurrent evaluation is a specific feature of the processing. Experimental data of by-line control are presented. The results obtained have shown the applicability of automatic signal processing to the testing under industrial conditions, which would permit to increase the testing efficiency to eliminate the subjectivism in assessment of testing results and to improve working conditions

  7. Applying dynamic data collection to improve dry electrode system performance for a P300-based brain-computer interface

    Science.gov (United States)

    Clements, J. M.; Sellers, E. W.; Ryan, D. B.; Caves, K.; Collins, L. M.; Throckmorton, C. S.

    2016-12-01

    Objective. Dry electrodes have an advantage over gel-based ‘wet’ electrodes by providing quicker set-up time for electroencephalography recording; however, the potentially poorer contact can result in noisier recordings. We examine the impact that this may have on brain-computer interface communication and potential approaches for mitigation. Approach. We present a performance comparison of wet and dry electrodes for use with the P300 speller system in both healthy participants and participants with communication disabilities (ALS and PLS), and investigate the potential for a data-driven dynamic data collection algorithm to compensate for the lower signal-to-noise ratio (SNR) in dry systems. Main results. Performance results from sixteen healthy participants obtained in the standard static data collection environment demonstrate a substantial loss in accuracy with the dry system. Using a dynamic stopping algorithm, performance may have been improved by collecting more data in the dry system for ten healthy participants and eight participants with communication disabilities; however, the algorithm did not fully compensate for the lower SNR of the dry system. An analysis of the wet and dry system recordings revealed that delta and theta frequency band power (0.1-4 Hz and 4-8 Hz, respectively) are consistently higher in dry system recordings across participants, indicating that transient and drift artifacts may be an issue for dry systems. Significance. Using dry electrodes is desirable for reduced set-up time; however, this study demonstrates that online performance is significantly poorer than for wet electrodes for users with and without disabilities. We test a new application of dynamic stopping algorithms to compensate for poorer SNR. Dynamic stopping improved dry system performance; however, further signal processing efforts are likely necessary for full mitigation.

  8. Automatic focusing system of BSST in Antarctic

    Science.gov (United States)

    Tang, Peng-Yi; Liu, Jia-Jing; Zhang, Guang-yu; Wang, Jian

    2015-10-01

    Automatic focusing (AF) technology plays an important role in modern astronomical telescopes. Based on the focusing requirement of BSST (Bright Star Survey Telescope) in Antarctic, an AF system is set up. In this design, functions in OpenCV is used to find stars, the algorithm of area, HFD or FWHM are used to degree the focus metric by choosing. Curve fitting method is used to find focus position as the method of camera moving. All these design are suitable for unattended small telescope.

  9. Automatic Visualization of Software Requirements: Reactive Systems

    International Nuclear Information System (INIS)

    Castello, R.; Mili, R.; Tollis, I.G.; Winter, V.

    1999-01-01

    In this paper we present an approach that facilitates the validation of high consequence system requirements. This approach consists of automatically generating a graphical representation from an informal document. Our choice of a graphical notation is statecharts. We proceed in two steps: we first extract a hierarchical decomposition tree from a textual description, then we draw a graph that models the statechart in a hierarchical fashion. The resulting drawing is an effective requirements assessment tool that allows the end user to easily pinpoint inconsistencies and incompleteness

  10. (a,k)-Anonymous Scheme for Privacy-Preserving Data Collection in IoT-based Healthcare Services Systems.

    Science.gov (United States)

    Li, Hongtao; Guo, Feng; Zhang, Wenyin; Wang, Jie; Xing, Jinsheng

    2018-02-14

    The widely use of IoT technologies in healthcare services has pushed forward medical intelligence level of services. However, it also brings potential privacy threat to the data collection. In healthcare services system, health and medical data that contains privacy information are often transmitted among networks, and such privacy information should be protected. Therefore, there is a need for privacy-preserving data collection (PPDC) scheme to protect clients (patients) data. We adopt (a,k)-anonymity model as privacy pretection scheme for data collection, and propose a novel anonymity-based PPDC method for healthcare services in this paper. The threat model is analyzed in the client-server-to-user (CS2U) model. On client-side, we utilize (a,k)-anonymity notion to generate anonymous tuples which can resist possible attack, and adopt a bottom-up clustering method to create clusters that satisfy a base privacy level of (a 1 ,k 1 )-anonymity. On server-side, we reduce the communication cost through generalization technology, and compress (a 1 ,k 1 )-anonymous data through an UPGMA-based cluster combination method to make the data meet the deeper level of privacy (a 2 ,k 2 )-anonymity (a 1  ≥ a 2 , k 2  ≥ k 1 ). Theoretical analysis and experimental results prove that our scheme is effective in privacy-preserving and data quality.

  11. Automatic system for detecting pornographic images

    Science.gov (United States)

    Ho, Kevin I. C.; Chen, Tung-Shou; Ho, Jun-Der

    2002-09-01

    Due to the dramatic growth of network and multimedia technology, people can more easily get variant information by using Internet. Unfortunately, it also makes the diffusion of illegal and harmful content much easier. So, it becomes an important topic for the Internet society to protect and safeguard Internet users from these content that may be encountered while surfing on the Net, especially children. Among these content, porno graphs cause more serious harm. Therefore, in this study, we propose an automatic system to detect still colour porno graphs. Starting from this result, we plan to develop an automatic system to search porno graphs or to filter porno graphs. Almost all the porno graphs possess one common characteristic that is the ratio of the size of skin region and non-skin region is high. Based on this characteristic, our system first converts the colour space from RGB colour space to HSV colour space so as to segment all the possible skin-colour regions from scene background. We also apply the texture analysis on the selected skin-colour regions to separate the skin regions from non-skin regions. Then, we try to group the adjacent pixels located in skin regions. If the ratio is over a given threshold, we can tell if the given image is a possible porno graph. Based on our experiment, less than 10% of non-porno graphs are classified as pornography, and over 80% of the most harmful porno graphs are classified correctly.

  12. Woefzela - An open-source platform for ASR data collection in the developing world

    CSIR Research Space (South Africa)

    De Vries, NJ

    2011-08-01

    Full Text Available to facilitate the efficient collection of speech data for Automatic Speech Recognition system development. The tool was designed for use in typical developing-world conditions; they present the relevant design choices and analyse the effectiveness of this tool...

  13. Automatic Generation of Supervisory Control System Software Using Graph Composition

    Science.gov (United States)

    Nakata, Hideo; Sano, Tatsuro; Kojima, Taizo; Seo, Kazuo; Uchida, Tomoyuki; Nakamura, Yasuaki

    This paper describes the automatic generation of system descriptions for SCADA (Supervisory Control And Data Acquisition) systems. The proposed method produces various types of data and programs for SCADA systems from equipment definitions using conversion rules. At first, this method makes directed graphs, which represent connections between the equipment, from equipment definitions. System descriptions are generated using the conversion rules, by analyzing these directed graphs, and finding the groups of equipment that involve similar operations. This method can make the conversion rules multi levels by using the composition of graphs, and can reduce the number of rules. The developer can define and manage these rules efficiently.

  14. AUTOMATIC RAILWAY POWER LINE EXTRACTION USING MOBILE LASER SCANNING DATA

    Directory of Open Access Journals (Sweden)

    S. Zhang

    2016-06-01

    Full Text Available Research on power line extraction technology using mobile laser point clouds has important practical significance on railway power lines patrol work. In this paper, we presents a new method for automatic extracting railway power line from MLS (Mobile Laser Scanning data. Firstly, according to the spatial structure characteristics of power-line and trajectory, the significant data is segmented piecewise. Then, use the self-adaptive space region growing method to extract power lines parallel with rails. Finally use PCA (Principal Components Analysis combine with information entropy theory method to judge a section of the power line whether is junction or not and which type of junction it belongs to. The least squares fitting algorithm is introduced to model the power line. An evaluation of the proposed method over a complicated railway point clouds acquired by a RIEGL VMX450 MLS system shows that the proposed method is promising.

  15. AUTOMATIC ROAD SIGN INVENTORY USING MOBILE MAPPING SYSTEMS

    Directory of Open Access Journals (Sweden)

    M. Soilán

    2016-06-01

    Full Text Available The periodic inspection of certain infrastructure features plays a key role for road network safety and preservation, and for developing optimal maintenance planning that minimize the life-cycle cost of the inspected features. Mobile Mapping Systems (MMS use laser scanner technology in order to collect dense and precise three-dimensional point clouds that gather both geometric and radiometric information of the road network. Furthermore, time-stamped RGB imagery that is synchronized with the MMS trajectory is also available. In this paper a methodology for the automatic detection and classification of road signs from point cloud and imagery data provided by a LYNX Mobile Mapper System is presented. First, road signs are detected in the point cloud. Subsequently, the inventory is enriched with geometrical and contextual data such as orientation or distance to the trajectory. Finally, semantic content is given to the detected road signs. As point cloud resolution is insufficient, RGB imagery is used projecting the 3D points in the corresponding images and analysing the RGB data within the bounding box defined by the projected points. The methodology was tested in urban and road environments in Spain, obtaining global recall results greater than 95%, and F-score greater than 90%. In this way, inventory data is obtained in a fast, reliable manner, and it can be applied to improve the maintenance planning of the road network, or to feed a Spatial Information System (SIS, thus, road sign information can be available to be used in a Smart City context.

  16. Big Data technology in traffic: A case study of automatic counters

    Directory of Open Access Journals (Sweden)

    Janković Slađana R.

    2016-01-01

    Full Text Available Modern information and communication technologies together with intelligent devices provide a continuous inflow of large amounts of data that are used by traffic and transport systems. Collecting traffic data does not represent a challenge nowadays, but the issues remains in relation to storing and processing increasing amounts of data. In this paper we have investigated the possibilities of using Big Data technology to store and process data in the transport domain. The term Big Data refers to a large volume of information resource, its velocity and variety, far beyond the capabilities of commonly used software for storing, processing and data management. In our case study, Apache™ Hadoop® Big Data was used for processing data collected from 10 automatic traffic counters set up in Novi Sad and its surroundings. Indicators of traffic load which were calculated using the Big Data platforms were presented using tables and graphs in Microsoft Office Excel tool. The visualization and geolocation of the obtained indicators were performed using the Microsoft Business Intelligence (BI tools such as: Excel Power View and Excel Power Map. This case study showed that Big Data technologies combined with the BI tools can be used as a reliable support in monitoring of the traffic management systems.

  17. Feasibility and desirability study of implementing a duty-cycle data-collection system as part of NPRDS

    International Nuclear Information System (INIS)

    Van Howe, K.R.; Koppe, R.H.; Voegtle, R.B.; Kline, S.C.; Olson, E.A.J.

    The objective of this project was to investigate cost-effective ways to improve the Nuclear Plant Reliability Data System (NPRDS) failure statistics as they are affected by the component-usage question. The nominal way to improve these statistics is to record, and to some extent describe, each operation of a safety component as well as each failure. Failures per demand or failures per operating hour could then be determined. It was recognized that such component-usage data collection for even a small number of key components could be costly or otherwise impractical. The reporting requirements for nuclear units are already burdensome and frequently redundant or overlapping. Therefore, the desirability of these information retrieval plans, in terms of the expected burdens and benefits, was also a consideration to be addressed, in addition to the feasibility of actually implementing such data collection

  18. Development of automatic techniques for GPS data management

    International Nuclear Information System (INIS)

    Park, Pil Ho

    2001-06-01

    It is necessary for GPS center to establish automatization as effective management of GPS network including data gathering, data transformation, data backup, data sending to IGS (International GPS Service for geodynamics), and precise ephemerides gathering. The operating program of GPS center has been adopted at KCSC (Korea Cadastral Survey Corporation), NGI (National Geography Institute), MOMAF (Ministry of Maritime Affairs and Fisheries) without self-development of core technique. The automatic management of GPS network is consists of GPS data management and data processing. It is also fundamental technique, which should be accomplished by every GPS centers. Therefore, this study carried out analyzing of Japanese GPS center, which has accomplished automatization by module considering applicability for domestic GPS centers

  19. Microprocessor-controlled system for automatic acquisition of potentiometric data and their non-linear least-squares fit in equilibrium studies.

    Science.gov (United States)

    Gampp, H; Maeder, M; Zuberbühler, A D; Kaden, T A

    1980-06-01

    A microprocessor-controlled potentiometric titration apparatus for equilibrium studies is described. The microprocessor controls the stepwise addition of reagent, monitors the pH until it becomes constant and stores the constant value. The data are recorded on magnetic tape by a cassette recorder with an RS232 input-output interface. A non-linear least-squares program based on Marquardt's modification of the Newton-Gauss method is discussed and its performance in the calculation of equilibrium constants is exemplified. An HP 9821 desk-top computer accepts the data from the magnetic tape recorder. In addition to a fully automatic fitting procedure, the program allows manual adjustment of the parameters. Three examples are discussed with regard to performance and reproducibility.

  20. A Machine Vision System for Automatically Grading Hardwood Lumber - (Industrial Metrology)

    Science.gov (United States)

    Richard W. Conners; Tai-Hoon Cho; Chong T. Ng; Thomas T. Drayer; Philip A. Araman; Robert L. Brisbon

    1992-01-01

    Any automatic system for grading hardwood lumber can conceptually be divided into two components. One of these is a machine vision system for locating and identifying grading defects. The other is an automatic grading program that accepts as input the output of the machine vision system and, based on these data, determines the grade of a board. The progress that has...

  1. Collective Analysis of Qualitative Data

    DEFF Research Database (Denmark)

    Simonsen, Jesper; Friberg, Karin

    2014-01-01

    What. Many students and practitioners do not know how to systematically process qualitative data once it is gathered—at least not as a collective effort. This chapter presents two workshop techniques, affinity diagramming and diagnostic mapping, that support collective analysis of large amounts...... of qualitative data. Affinity diagramming is used to make collective analysis and interpretations of qualitative data to identify core problems that need to be addressed in the design process. Diagnostic mapping supports collective interpretation and description of these problems and how to intervene in them. We....... In particular, collective analysis can be used to identify, understand, and act on complex design problems that emerge, for example, after the introduction of new tech- nologies. Such problems might be hard to clarify, and the basis for the analysis often involves large amounts of unstructured qualitative data...

  2. Wireless sensor network-based greenhouse environment monitoring and automatic control system for dew condensation prevention.

    Science.gov (United States)

    Park, Dae-Heon; Park, Jang-Woo

    2011-01-01

    Dew condensation on the leaf surface of greenhouse crops can promote diseases caused by fungus and bacteria, affecting the growth of the crops. In this paper, we present a WSN (Wireless Sensor Network)-based automatic monitoring system to prevent dew condensation in a greenhouse environment. The system is composed of sensor nodes for collecting data, base nodes for processing collected data, relay nodes for driving devices for adjusting the environment inside greenhouse and an environment server for data storage and processing. Using the Barenbrug formula for calculating the dew point on the leaves, this system is realized to prevent dew condensation phenomena on the crop's surface acting as an important element for prevention of diseases infections. We also constructed a physical model resembling the typical greenhouse in order to verify the performance of our system with regard to dew condensation control.

  3. Wireless Sensor Network-Based Greenhouse Environment Monitoring and Automatic Control System for Dew Condensation Prevention

    Science.gov (United States)

    Park, Dae-Heon; Park, Jang-Woo

    2011-01-01

    Dew condensation on the leaf surface of greenhouse crops can promote diseases caused by fungus and bacteria, affecting the growth of the crops. In this paper, we present a WSN (Wireless Sensor Network)-based automatic monitoring system to prevent dew condensation in a greenhouse environment. The system is composed of sensor nodes for collecting data, base nodes for processing collected data, relay nodes for driving devices for adjusting the environment inside greenhouse and an environment server for data storage and processing. Using the Barenbrug formula for calculating the dew point on the leaves, this system is realized to prevent dew condensation phenomena on the crop’s surface acting as an important element for prevention of diseases infections. We also constructed a physical model resembling the typical greenhouse in order to verify the performance of our system with regard to dew condensation control. PMID:22163813

  4. Making sense of the shadows: priorities for creating a learning healthcare system based on routinely collected data.

    Science.gov (United States)

    Deeny, Sarah R; Steventon, Adam

    2015-08-01

    Socrates described a group of people chained up inside a cave, who mistook shadows of objects on a wall for reality. This allegory comes to mind when considering 'routinely collected data'-the massive data sets, generated as part of the routine operation of the modern healthcare service. There is keen interest in routine data and the seemingly comprehensive view of healthcare they offer, and we outline a number of examples in which they were used successfully, including the Birmingham OwnHealth study, in which routine data were used with matched control groups to assess the effect of telephone health coaching on hospital utilisation.Routine data differ from data collected primarily for the purposes of research, and this means that analysts cannot assume that they provide the full or accurate clinical picture, let alone a full description of the health of the population. We show that major methodological challenges in using routine data arise from the difficulty of understanding the gap between patient and their 'data shadow'. Strategies to overcome this challenge include more extensive data linkage, developing analytical methods and collecting more data on a routine basis, including from the patient while away from the clinic. In addition, creating a learning health system will require greater alignment between the analysis and the decisions that will be taken; between analysts and people interested in quality improvement; and between the analysis undertaken and public attitudes regarding appropriate use of data. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

  5. Permanent automatic recalibration system for scintillation camera

    International Nuclear Information System (INIS)

    Auphan, Michel.

    1974-01-01

    A permanent automatic recalibration system for a scintillation camera, of the type consisting chiefly of a collimator if necessary, a scintillator, a light guide and a network of n photomultipliers coupled to a display system, is described. It uses a device to form a single reference light signal common to all the photomultiplication lines, integrated to these latter and associated with a periodic calibration control generator. By means of associated circuits governed by the control generator the gain in each line is brought to and/or maintained at a value between fixed upper and lower limits. Steps are taken so that any gain variation in a given line is adjusted with respect to the reference light signal common to all the lines. The light signal falls preferably in the same part of the spectrum as the scintillations formed in the scintillator [fr

  6. Microprocessor controlled system for automatic and semi-automatic syntheses of radiopharmaceuticals

    International Nuclear Information System (INIS)

    Ruth, T.J.; Adam, M.J.; Morris, D.; Jivan, S.

    1986-01-01

    A computer based system has been constructed to control the automatic synthesis of 2-deoxy-2-( 18 F)fluoro-D-glucose and is also being used in the development of an automatic synthesis of L-6-( 18 F)fluorodopa. (author)

  7. Fuzzy-Neural Automatic Daylight Control System

    Directory of Open Access Journals (Sweden)

    Grif H. Şt.

    2011-12-01

    Full Text Available The paper presents the design and the tuning of a CMAC controller (Cerebellar Model Articulation Controller implemented in an automatic daylight control application. After the tuning process of the controller, the authors studied the behavior of the automatic lighting control system (ALCS in the presence of luminance disturbances. The luminance disturbances were produced by the authors in night conditions and day conditions as well. During the night conditions, the luminance disturbances were produced by turning on and off a halogen desk lamp. During the day conditions the luminance disturbances were produced in two ways: by daylight contributions changes achieved by covering and uncovering a part of the office window and by turning on and off a halogen desk lamp. During the day conditions the luminance disturbances, produced by turning on and off the halogen lamp, have a smaller amplitude than those produced during the night conditions. The luminance disturbance during the night conditions was a helpful tool to select the proper values of the learning rate for CMAC controller. The luminance disturbances during the day conditions were a helpful tool to demonstrate the right setting of the CMAC controller.

  8. THE NOISE IMMUNITY OF THE DIGITAL DEMODULATOR MFM-AM SIGNAL USED IN DATA COMMUNICATIONS SYSTEMS OF AIR TRAFFIC CONTROL WITH AUTOMATIC DEPENDENT SURVEILLANCE AGAINST A NON-GAUSSIAN NOISE

    Directory of Open Access Journals (Sweden)

    A. L. Senyavskiy

    2015-01-01

    Full Text Available The article analyzes the robustness of the digital demodulator of the signal with the lowest frequency shift keying at a subcarrier frequency with respect to non-Gaussian interference type of atmospheric, industrial noise and interfering frequency -and phase-shift keyed signals. This type of demodulator is used for the transmission of navigation data in the systems of air traffic control with automatic dependent surveillance.

  9. Automatic Data Traffic Control on DSM Architecture

    Science.gov (United States)

    Frumkin, Michael; Jin, Hao-Qiang; Yan, Jerry; Kwak, Dochan (Technical Monitor)

    2000-01-01

    We study data traffic on distributed shared memory machines and conclude that data placement and grouping improve performance of scientific codes. We present several methods which user can employ to improve data traffic in his code. We report on implementation of a tool which detects the code fragments causing data congestions and advises user on improvements of data routing in these fragments. The capabilities of the tool include deduction of data alignment and affinity from the source code; detection of the code constructs having abnormally high cache or TLB misses; generation of data placement constructs. We demonstrate the capabilities of the tool on experiments with NAS parallel benchmarks and with a simple computational fluid dynamics application ARC3D.

  10. Computer applications: Automatic control system for high-voltage accelerator

    International Nuclear Information System (INIS)

    Bryukhanov, A.N.; Komissarov, P.Yu.; Lapin, V.V.; Latushkin, S.T.. Fomenko, D.E.; Yudin, L.I.

    1992-01-01

    An automatic control system for a high-voltage electrostatic accelerator with an accelerating potential of up to 500 kV is described. The electronic apparatus on the high-voltage platform is controlled and monitored by means of a fiber-optic data-exchange system. The system is based on CAMAC modules that are controlled by a microprocessor crate controller. Data on accelerator operation are represented and control instructions are issued by means of an alphanumeric terminal. 8 refs., 6 figs

  11. [Automatic Extraction and Analysis of Dosimetry Data in Radiotherapy Plans].

    Science.gov (United States)

    Song, Wei; Zhao, Di; Lu, Hong; Zhang, Biyun; Ma, Jun; Yu, Dahai

    To improve the efficiency and accuracy of extraction and analysis of dosimetry data in radiotherapy plans for a batch of patients. With the interface function provided in Matlab platform, a program was written to extract the dosimetry data exported from treatment planning system in DICOM RT format and exported the dose-volume data to an Excel file with the SPSS compatible format. This method was compared with manual operation for 14 gastric carcinoma patients to validate the efficiency and accuracy. The output Excel data were compatible with SPSS in format, the dosimetry data error for PTV dose interval of 90%-98%, PTV dose interval of 99%-106% and all OARs were -3.48E-5 ± 3.01E-5, -1.11E-3 ± 7.68E-4, -7.85E-5 ± 9.91E-5 respectively. Compared with manual operation, the time required was reduced from 5.3 h to 0.19 h and input error was reduced from 0.002 to 0. The automatic extraction of dosimetry data in DICOM RT format for batch patients, the SPSS compatible data exportation, quick analysis were achieved in this paper. The efficiency of clinical researches based on dosimetry data analysis of large number of patients will be improved with this methods.

  12. Development of inspection data collection and evaluation system for large scale MOX fuel fabrication plant safeguards (3)

    International Nuclear Information System (INIS)

    Kumakura, Shinichi; Masuda, Shoichiro; Iso, Shoko; Hisamatsu, Yoshinori; Kurobe, Hiroko; Nakajima, Shinji

    2015-01-01

    Inspection Data Collection and Evaluation System is the system to store inspection data and operator declaration data collected from various measurement equipment, which is installed in fuel fabrication processes of the large-scale MOX fuel fabrication plant, and to make safeguards evaluation based on Near Real Time Accountancy (NRTA) using these data. Nuclear Material Control Center developed the simulator to simulate fuel fabrication process, in-process material inventory/flow data and the measurement data and the adequacy/impact to the uncertainty of the material balance using the simulation results, such as the facility operation and the operational status, has been reviewed. Following the 34th INMM Japan chapter presentation, the model similar to the real nuclear material accountancy during the fuel fabrication process was simulated and the nuclear material accountancy and its uncertainty (Sigma MUF) have been reviewed. Some findings have been obtained, such as regarding evaluation related indicators for verification under a more realistic accountancy which could be applied by operator. (author)

  13. An automatic injection system for rapid radiochemistry

    International Nuclear Information System (INIS)

    Nurmia, M.J.; Kreek, S.A.; Kadkhodayan, B.; Gregorich, K.E.; Lee, D.M.; Hoffman, D.C.

    1992-01-01

    A description is given of the Automated Injection System (AIS), a pneumatically actuated device for automated collection of nuclear reaction products from a He/KCl gas jet transport system. The AIS is used with the Automated Chemical Chromatographic Element Separation System; together these two devices facilitate completely automated separation procedures with improved speed and reproducibility

  14. Global Positioning System data collection, processing, and analysis conducted by the U.S. Geological Survey Earthquake Hazards Program

    Science.gov (United States)

    Murray, Jessica R.; Svarc, Jerry L.

    2017-01-01

    The U.S. Geological Survey Earthquake Science Center collects and processes Global Positioning System (GPS) data throughout the western United States to measure crustal deformation related to earthquakes and tectonic processes as part of a long‐term program of research and monitoring. Here, we outline data collection procedures and present the GPS dataset built through repeated temporary deployments since 1992. This dataset consists of observations at ∼1950 locations. In addition, this article details our data processing and analysis procedures, which consist of the following. We process the raw data collected through temporary deployments, in addition to data from continuously operating western U.S. GPS stations operated by multiple agencies, using the GIPSY software package to obtain position time series. Subsequently, we align the positions to a common reference frame, determine the optimal parameters for a temporally correlated noise model, and apply this noise model when carrying out time‐series analysis to derive deformation measures, including constant interseismic velocities, coseismic offsets, and transient postseismic motion.

  15. An overview of future EU health systems. An insight into governance, primary care, data collection and citizens' participation.

    Science.gov (United States)

    Quaglio, Gianluca; Figueras, Josep; Mantoan, Domenico; Dawood, Amr; Karapiperis, Theodoros; Costongs, Caroline; Bernal-Delgado, Enrique

    2018-03-26

    Health systems in the European Union (EU) are being questioned over their effectiveness and sustainability. In pursuing both goals, they have to conciliate coexisting, not always aligned, realities. This paper originated from a workshop entitled 'Health systems for the future' held at the European Parliament. Experts and decision makers were asked to discuss measures that may increase the effectiveness and sustainability of health systems, namely: (i) increasing citizens' participation; (ii) the importance of primary care in providing integrated services; (iii) improving the governance and (iv) fostering better data collection and information channels to support the decision making process. In the parliamentary debate, was discussed the concept that, in the near future, health systems' effectiveness and sustainability will very much depend on effective access to integrated services where primary care is pivotal, a clearer shift from care-oriented systems to health promotion and prevention, a profound commitment to good governance, particularly to stakeholders participation, and a systematic reuse of data meant to build health data-driven learning systems. Many health issues, such as future health systems in the EU, are potentially transformative and hence an intense political issue. It is policy-making leadership that will mostly determine how well EU health systems are prepared to face future challenges.

  16. Automatic data acquisition of anthropological measurements

    DEFF Research Database (Denmark)

    Lynnerup, N; Lynnerup, O

    1993-01-01

    A computer program in BASIC is presented which enables the input of measurement data from a caliper directly into specific records in a dBASE IV or PARADOX database. The program circumvents the tedious procedure of first recording measurement data manually and then entering the data into a computer....... Thus much time can be saved and the risk of wrong data entry is lowered. The program was easy to use, and no significant problems were encountered. Necessary hardware is a standard IBM compatible desktop computer, Mitotoyu Digimatic (TM) calipers and a Mitotoyu Digimatic MUX-10 Multiplexer (TM)....

  17. Nuclear data collection on virtual instrument

    International Nuclear Information System (INIS)

    Yang Qiang; Lai Wanchang; Hua Yongtai

    2007-01-01

    This paper introduced a kind of nuclear data collection system using a technique of mixed programming. The system uses parallel port to transport data, MATLAB signal processing software is used for data analysis. It has significantly reduced the difficult of spectral analysis software by using the technique. It has realized signal acquisition in real time and data analysis. (authors)

  18. Tracking of nuclear shipments with automatic vehicle location systems

    International Nuclear Information System (INIS)

    Colhoun, C.J.K.

    1989-01-01

    A complete Automatic Vehicle Location System (AVL) consists of three main elements: (1) the location sensor in the vehicle, this device constantly determines the coordinates of the vehicles position; (2) the radio link between vehicle and central base; (3) the data processing and display in the central base. For all three elements there are several solutions. The optimal combination of the different techniques depends on the requirements of the special application

  19. The TS 600: automatic control system for eddy currents

    International Nuclear Information System (INIS)

    Poulet, J.P.

    1986-10-01

    In the scope of fabrication and in service inspection of the PWR steam generator tubing bendle, FRAMATOME developed an automatic Eddy Current testing system: TS600. Based on a mini-computer, TS600 allows to digitize, to store and to process data in various ways, so it is possible to perform several kinds of inspection: conventional inservice inspection, roll area profilometry...... TS600 can also be used to develop new methods of examination [fr

  20. Data-driven automatic parking constrained control for four-wheeled mobile vehicles

    OpenAIRE

    Wenxu Yan; Jing Deng; Dezhi Xu

    2016-01-01

    In this article, a novel data-driven constrained control scheme is proposed for automatic parking systems. The design of the proposed scheme only depends on the steering angle and the orientation angle of the car, and it does not involve any model information of the car. Therefore, the proposed scheme-based automatic parking system is applicable to different kinds of cars. In order to further reduce the desired trajectory coordinate tracking errors, a coordinates compensation algorithm is als...

  1. Quality assurance for screening mammography data collection systems in 22 countries.

    NARCIS (Netherlands)

    Klabunde, C.N.; Sancho-Garnier, H.; Broeders, M.E.A.C.; Thoresen, S.; Rodrigues, V.J.; Ballard-Barbash, R.

    2001-01-01

    OBJECTIVES: To document the mammography data that are gathered by the organized screening programs participating in the International Breast Cancer Screening Network (IBSN), the nature of their procedures for data quality assurance, and the measures used to assess program performance and impact.

  2. Automatic delimitation of microwatershed using SRTM data of the NASA

    Directory of Open Access Journals (Sweden)

    Freddy Aníbal Jumbo Castillo

    2015-12-01

    Full Text Available The watershed as the basic territorial unit of planning and management of water resources, requires its proper delimitation of the catchment or drainage area, faced with this situation, the lack of geographic information of Casacay river micro watersheds, hydrographic unit should be resolved, for this purpose the research was aimed at automatic delimitation of micro watersheds using of Geographic Information Systems (GIS techniques and the project Shuttle Radar Topographic Mission (SRTM 30 meters spatial resolution data. The selected methodology was the Pfafstetter one, with which nine micro watersheds were obtained with their respective codification allowing to continue with watersheds standardization adopted by Ecuador Water's Secretariat. With the investigation results watersheds will be updated with more detail information, promoting the execution of tasks or activities related to the integrated management of the hydrographic unit studied

  3. A Mobile Health Data Collection System for Remote Areas to Monitor Women Participating in a Cervical Cancer Screening Campaign.

    Science.gov (United States)

    Quercia, Kelly; Tran, Phuong Lien; Jinoro, Jéromine; Herniainasolo, Joséa Lea; Viviano, Manuela; Vassilakos, Pierre; Benski, Caroline; Petignat, Patrick

    2018-04-01

    Barriers to efficient cervical cancer screening in low- and medium-income countries include the lack of systematic monitoring of the participants' data. The aim of this study was to assess the feasibility of a mobile health (m-Health) data collection system to facilitate monitoring of women participating to cervical cancer screening campaign. Women aged 30-65 years, participating in a cervical cancer screening campaign in Ambanja, Madagascar, were invited to participate in the study. Cervical Cancer Prevention System, an m-Health application, allows the registration of clinical data, while women are undergoing cervical cancer screening. All data registered in the smartphone were transmitted onto a secure, Web-based platform through the use of an Internet connection. Healthcare providers had access to the central database and could use it for the follow-up visits. Quality of data was assessed by computing the percentage of key data missing. A total of 151 women were recruited in the study. Mean age of participants was 41.8 years. The percentage of missing data for the key variables was less than 0.02%, corresponding to one woman's medical history data, which was not sent to the central database. Technical problems, including transmission of photos, human papillomavirus test results, and pelvic examination data, have subsequently been solved through a system update. The quality of the data was satisfactory and allowed monitoring of cervical cancer screening data of participants. Larger studies evaluating the efficacy of the system for the women's follow-up are needed in order to confirm its efficiency on a long-term scale.

  4. Geographic Names Information System (GNIS) - USGS National Map Downloadable Data Collection

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — The Geographic Names Information System (GNIS) is the Federal standard for geographic nomenclature. The U.S. Geological Survey developed the GNIS for the U.S. Board...

  5. MEMOPS: data modelling and automatic code generation.

    Science.gov (United States)

    Fogh, Rasmus H; Boucher, Wayne; Ionides, John M C; Vranken, Wim F; Stevens, Tim J; Laue, Ernest D

    2010-03-25

    In recent years the amount of biological data has exploded to the point where much useful information can only be extracted by complex computational analyses. Such analyses are greatly facilitated by metadata standards, both in terms of the ability to compare data originating from different sources, and in terms of exchanging data in standard forms, e.g. when running processes on a distributed computing infrastructure. However, standards thrive on stability whereas science tends to constantly move, with new methods being developed and old ones modified. Therefore maintaining both metadata standards, and all the code that is required to make them useful, is a non-trivial problem. Memops is a framework that uses an abstract definition of the metadata (described in UML) to generate internal data structures and subroutine libraries for data access (application programming interfaces--APIs--currently in Python, C and Java) and data storage (in XML files or databases). For the individual project these libraries obviate the need for writing code for input parsing, validity checking or output. Memops also ensures that the code is always internally consistent, massively reducing the need for code reorganisation. Across a scientific domain a Memops-supported data model makes it easier to support complex standards that can capture all the data produced in a scientific area, share them among all programs in a complex software pipeline, and carry them forward to deposition in an archive. The principles behind the Memops generation code will be presented, along with example applications in Nuclear Magnetic Resonance (NMR) spectroscopy and structural biology.

  6. EPA Linked Open Data (Collection)

    Data.gov (United States)

    U.S. Environmental Protection Agency — This is a collection item referencing the following EPA Linked Data resources: - EPA Facility Registry Service (FRS) - EPA Substance Registry Service (SRS) -...

  7. Water Column Sonar Data Collection

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The collection and analysis of water column sonar data is a relatively new avenue of research into the marine environment. Primary uses include assessing biological...

  8. Distributed privacy preserving data collection

    KAUST Repository

    Xue, Mingqiang

    2011-01-01

    We study the distributed privacy preserving data collection problem: an untrusted data collector (e.g., a medical research institute) wishes to collect data (e.g., medical records) from a group of respondents (e.g., patients). Each respondent owns a multi-attributed record which contains both non-sensitive (e.g., quasi-identifiers) and sensitive information (e.g., a particular disease), and submits it to the data collector. Assuming T is the table formed by all the respondent data records, we say that the data collection process is privacy preserving if it allows the data collector to obtain a k-anonymized or l-diversified version of T without revealing the original records to the adversary. We propose a distributed data collection protocol that outputs an anonymized table by generalization of quasi-identifier attributes. The protocol employs cryptographic techniques such as homomorphic encryption, private information retrieval and secure multiparty computation to ensure the privacy goal in the process of data collection. Meanwhile, the protocol is designed to leak limited but non-critical information to achieve practicability and efficiency. Experiments show that the utility of the anonymized table derived by our protocol is in par with the utility achieved by traditional anonymization techniques. © 2011 Springer-Verlag.

  9. Travel time data collection for measurement of advanced traveler information systems accuracy

    Science.gov (United States)

    2003-06-01

    The objective of this white paper is to recommend an approach to measuring ATIS travel time accuracy so that ITS planners might have the data they need to make cost effective decisions regarding deployment of surveillance technologies to support ATIS...

  10. Automatic charge control system for satellites

    Science.gov (United States)

    Shuman, B. M.; Cohen, H. A.

    1985-01-01

    The SCATHA and the ATS-5 and 6 spacecraft provided insights to the problem of spacecraft charging at geosychronous altitudes. Reduction of the levels of both absolute and differential charging was indicated, by the emission of low energy neutral plasma. It is appropriate to complete the transition from experimental results to the development of a system that will sense the state-of-charge of a spacecraft, and, when a predetermined threshold is reached, will respond automatically to reduce it. A development program was initiated utilizing sensors comparable to the proton electrostatic analyzer, the surface potential monitor, and the transient pulse monitor that flew in SCATHA, and combine these outputs through a microprocessor controller to operate a rapid-start, low energy plasma source.

  11. Data collection system. Volume 1, Overview and operators manual; Volume 2, Maintenance manual; Appendices

    Energy Technology Data Exchange (ETDEWEB)

    Caudell, R.B.; Bauder, M.E.; Boyer, W.B.; French, R.E.; Isidoro, R.J.; Kaestner, P.C.; Perkins, W.G.

    1993-09-01

    Sandia National Laboratories (SNL) Instrumentation Development Department was tasked by the Defense Nuclear Agency (DNA) to record data on Tektronix RTD720 Digitizers on the HUNTERS TROPHY field test conducted at the Nevada Test Site (NTS) on September 18, 1992. This report contains a overview and description of the computer hardware and software that was used to acquire, reduce, and display the data. The document is divided into two volumes: an overview and operators manual (Volume 1) and a maintenance manual (Volume 2).

  12. Human-system Interfaces for Automatic Systems

    Energy Technology Data Exchange (ETDEWEB)

    OHara, J.M.; Higgins,J. (BNL); Fleger, S.; Barnes V. (NRC)

    2010-11-07

    Automation is ubiquitous in modern complex systems, and commercial nuclear- power plants are no exception. Automation is applied to a wide range of functions including monitoring and detection, situation assessment, response planning, and response implementation. Automation has become a 'team player' supporting personnel in nearly all aspects of system operation. In light of its increasing use and importance in new- and future-plants, guidance is needed to conduct safety reviews of the operator's interface with automation. The objective of this research was to develop such guidance. We first characterized the important HFE aspects of automation, including six dimensions: levels, functions, processes, modes, flexibility, and reliability. Next, we reviewed literature on the effects of all of these aspects of automation on human performance, and on the design of human-system interfaces (HSIs). Then, we used this technical basis established from the literature to identify general principles for human-automation interaction and to develop review guidelines. The guidelines consist of the following seven topics: automation displays, interaction and control, automation modes, automation levels, adaptive automation, error tolerance and failure management, and HSI integration. In addition, our study identified several topics for additional research.

  13. Platform attitude data acquisition system

    Digital Repository Service at National Institute of Oceanography (India)

    Afzulpurkar, S.

    A system for automatic acquisition of underwater platform attitude data has been designed, developed and tested in the laboratory. This is a micro controller based system interfacing dual axis inclinometer, high-resolution digital compass...

  14. TFTR data management system

    International Nuclear Information System (INIS)

    Randerson, L.; Chu, J.; Ludescher, C.; Malsbury, J.; Stark, W.

    1986-01-01

    Developments in the tokamak fusion test reactor (TFTR) data management system supporting data management system supporting data acquisition and off-line physics data reduction are described. Data from monitor points, timing channels, and transient recorder channels and other devices are acquired and stored for use by on-line tasks. Files are transferred off-line automatically. A configuration utility determines data acquired and files transferred. An event system driven by file arrival activates off-line reduction processes. A post-run process transfers files not shipped during runs. Files are archived to tape and are retrievable by digraph and shot number. Automatic skimming based on most recent access, file type, shot numbers, and user-set protection maintains the files required for post-run data reduction

  15. Ground penetrating radar and differential global positioning system data collected in April 2016 from Fire Island, New York

    Science.gov (United States)

    Forde, Arnell S.; Bernier, Julie C.; Miselis, Jennifer L.

    2018-02-22

    Researchers from the U.S. Geological Survey (USGS) conducted a long-term coastal morphologic-change study at Fire Island, New York, prior to and after Hurricane Sandy impacted the area in October 2012. The Fire Island Coastal Change project objectives include understanding the morphologic evolution of the barrier island system on a variety of time scales (months to centuries) and resolving storm-related impacts, post-storm beach response, and recovery. In April 2016, scientists from the USGS St. Petersburg Coastal and Marine Science Center conducted geophysical and sediment sampling surveys on Fire Island to characterize and quantify spatial variability in the subaerial geology with the goal of subsequently integrating onshore geology with other surf zone and nearshore datasets.  This report, along with the associated USGS data release, serves as an archive of ground penetrating radar (GPR) and post-processed differential global positioning system (DGPS) data collected from beach and back-barrier environments on Fire Island, April 6–13, 2016 (USGS Field Activity Number 2016-322-FA). Data products, including unprocessed GPR trace data, processed DGPS data, elevation-corrected subsurface profile images, geographic information system files, and accompanying Federal Geographic Data Committee metadata are available for download.

  16. An automatic monitoring system of leak current for testing TGC detectors based on LabVIEW

    International Nuclear Information System (INIS)

    Feng Cunfeng; Lu Taiguo; Yan Zhen; Wang Suojie; Zhu Chengguang; Sun Yansheng; He Mao

    2005-01-01

    An automatic monitoring system of leak current for testing TGC detectors with high voltage was set up by using the graphic LabVIEW platform and NI 4351 data acquisition card. The leak current was automatically monitored and recorded with this system, the time and the value of the leak current were showed instantly. Good efficiency and precision of monitoring were obtained. (authors)

  17. Collecting operational event data for statistical analysis

    International Nuclear Information System (INIS)

    Atwood, C.L.

    1994-09-01

    This report gives guidance for collecting operational data to be used for statistical analysis, especially analysis of event counts. It discusses how to define the purpose of the study, the unit (system, component, etc.) to be studied, events to be counted, and demand or exposure time. Examples are given of classification systems for events in the data sources. A checklist summarizes the essential steps in data collection for statistical analysis

  18. Collecting data in real time with postcards

    DEFF Research Database (Denmark)

    Yee, Kwang Chien; Kanstrup, Anne Marie; Bertelsen, Pernille

    2013-01-01

    Systems. These methods often involve cross-sectional, retrospective data collection. This paper describes the postcard method for prospective real-time data collection, both in paper format and electronic format. This paper then describes the results obtained using postcard techniques in Denmark...

  19. Interactive Visualization Systems and Data Integration Methods for Supporting Discovery in Collections of Scientific Information

    Science.gov (United States)

    2011-05-01

    projection for exploratory analysis. It also enables quantitative analysis. We show that this combination can be used to assisting users with the...cited in “Undiscovered Public Knowledge” Swanson succeeds in integrating Wilson’s ideas with Karl Popper’s critique of positivism from the 1934 “Logik...data that was performed in the prior studies. In this study, quantitative properties of the graph were used to identify records that merit

  20. Implementation of an electronic fingerprint-linked data collection system: a feasibility and acceptability study among Zambian female sex workers.

    Science.gov (United States)

    Wall, Kristin M; Kilembe, William; Inambao, Mubiana; Chen, Yi No; Mchoongo, Mwaka; Kimaru, Linda; Hammond, Yuna Tiffany; Sharkey, Tyronza; Malama, Kalonde; Fulton, T Roice; Tran, Alex; Halumamba, Hanzunga; Anderson, Sarah; Kishore, Nishant; Sarwar, Shawn; Finnegan, Trisha; Mark, David; Allen, Susan A

    2015-06-27

    Patient identification within and between health services is an operational challenge in many resource-limited settings. When following HIV risk groups for service provision and in the context of vaccine trials, patient misidentification can harm patient care and bias trial outcomes. Electronic fingerprinting has been proposed to identify patients over time and link patient data between health services. The objective of this study was to determine 1) the feasibility of implementing an electronic-fingerprint linked data capture system in Zambia and 2) the acceptability of this system among a key HIV risk group: female sex workers (FSWs). Working with Biometrac, a US-based company providing biometric-linked healthcare platforms, an electronic fingerprint-linked data capture system was developed for use by field recruiters among Zambian FSWs. We evaluated the technical feasibility of the system for use in the field in Zambia and conducted a pilot study to determine the acceptability of the system, as well as barriers to uptake, among FSWs. We found that implementation of an electronic fingerprint-linked patient tracking and data collection system was feasible in this relatively resource-limited setting (false fingerprint matching rate of 1/1000 and false rejection rate of lodges) could be addressed by recruiting women during less busy hours, in their own homes, in the presence of "Queen Mothers" (FSW organizers), or in the presence of a FSW that has already been fingerprinted. Our findings have major implications for key population research and improved health services provision. However, more work needs to be done to increase the acceptability of the electronic fingerprint-linked data capture system during field recruitment. This study indicated several potential avenues that will be explored to increase acceptability.

  1. Development of automatic editing system for MCNP library 'autonj'

    International Nuclear Information System (INIS)

    Maekawa, Fujio; Sakurai, Kiyoshi; Kume, Etsuo; Nomura, Yasushi; Kosako, Kazuaki; Kawasaki, Nobuo; Naito, Yoshitaka

    1999-12-01

    As an activity of the MCNP High-Temperature Library Production Working Group under the Nuclear Code Evaluation Special Committee of Nuclear Code Committee, the automatic editing system for MCNP library 'autonj' was developed. The autonj includes the NJOY-97 code as its main body, and is a system that enables us to easily produce cross section libraries for MCNP from evaluated nuclear data files such as JENDL-3.2. A temperature dependent library at six temperature points based on JENDL-3.2 was produced by using autonj. The autonj system and the temperature dependent library were installed on the JAERI AP3000 computer. (author)

  2. ASTRA - an automatic system for transport analysis in a tokamak

    International Nuclear Information System (INIS)

    Pereverzev, G.V.; Yushmanov, P.N.; Dnestrovskii, A.Yu.; Polevoi, A.R.; Tarasjan, K.N.; Zakharov, L.E.

    1991-08-01

    The set of codes described here - ASTRA (Automatic System of Transport Analysis) - is a flexible and effective tool for the study of transport mechanisms in reactor-oriented facilities of the tokamak type. Flexibility is provided within the ASTRA system by a wide choice of standard relationships, functions and subroutines representing various transport coefficients, methods of auxiliary heating and other physical processes in the tokamak plasma, as well as by the possibility of pre-setting transport equations and variables for data output in a simple and conseptually transparent form. The transport code produced by the ASTRA system provides an adequate representation of the discharges for present experimental conditions. (orig.)

  3. Automation of plasma-process fultext bibliography databases. An on-line data-collection, data-mining and data-input system

    International Nuclear Information System (INIS)

    Suzuki, Manabu; Pichl, Lukas; Murakami, Izumi; Kato, Takako; Sasaki, Akira

    2006-01-01

    Searching for relevant data, information retrieval, data extraction and data input are time- and resource-consuming activities in most data centers. Here we develop a Linux system automating the process in case of bibliography, abstract and fulltext databases. The present system is an open-source free-software low-cost solution that connects the target and provider databases in cyberspace through various web publishing formats. The abstract/fulltext relevance assessment is interfaced to external software modules. (author)

  4. 2013 International Conference on Mechatronics and Automatic Control Systems

    CERN Document Server

    2014-01-01

    This book examines mechatronics and automatic control systems. The book covers important emerging topics in signal processing, control theory, sensors, mechanic manufacturing systems and automation. The book presents papers from the 2013 International Conference on Mechatronics and Automatic Control Systems held in Hangzhou, China on August 10-11, 2013. .

  5. System for Automatic Generation of Examination Papers in Discrete Mathematics

    Science.gov (United States)

    Fridenfalk, Mikael

    2013-01-01

    A system was developed for automatic generation of problems and solutions for examinations in a university distance course in discrete mathematics and tested in a pilot experiment involving 200 students. Considering the success of such systems in the past, particularly including automatic assessment, it should not take long before such systems are…

  6. Automatic Generation of Symbolic Model for Parameterized Synchronous Systems

    Institute of Scientific and Technical Information of China (English)

    Wei-Wen Xu

    2004-01-01

    With the purpose of making the verification of parameterized system more general and easier, in this paper, a new and intuitive language PSL (Parameterized-system Specification Language) is proposed to specify a class of parameterized synchronous systems. From a PSL script, an automatic method is proposed to generate a constraint-based symbolic model. The model can concisely symbolically represent the collections of global states by counting the number of processes in a given state. Moreover, a theorem has been proved that there is a simulation relation between the original system and its symbolic model. Since the abstract and symbolic techniques are exploited in the symbolic model, state-explosion problem in traditional verification methods is efficiently avoided. Based on the proposed symbolic model, a reachability analysis procedure is implemented using ANSI C++ on UNIX platform. Thus, a complete tool for verifying the parameterized synchronous systems is obtained and tested for some cases. The experimental results show that the method is satisfactory.

  7. Automatic Analysis of Swift-XRT data

    Science.gov (United States)

    Evans, P. A.; Tyler, L. G.; Beardmore, A. P.; Osborne, J. P.

    2008-08-01

    The Swift spacecraft detects and autonomously observes ˜100 Gamma Ray Bursts (GRBs) per year, ˜96% of which are detected by the X-ray telescope (XRT). GRBs are accompanied by optical transients and the field of ground-based follow-up of GRBs has expanded significantly over the last few years, with rapid response instruments capable of responding to Swift triggers on timescales of minutes. To make the most efficient use of limited telescope time, follow-up astronomers need accurate positions of GRBs as soon as possible after the trigger. Additionally, information such as the X-ray light curve, is of interest when considering observing strategy. The Swift team at Leicester University have developed techniques to improve the accuracy of the GRB positions available from the XRT, and to produce science-grade X-ray light curves of GRBs. These techniques are fully automated, and are executed as soon as data are available.

  8. Radioactive Decay: Audio Data Collection

    Science.gov (United States)

    Struthers, Allan

    2009-01-01

    Many phenomena generate interesting audible time series. This data can be collected and processed using audio software. The free software package "Audacity" is used to demonstrate the process by recording, processing, and extracting click times from an inexpensive radiation detector. The high quality of the data is demonstrated with a simple…

  9. Collection of offshore human error probability data

    International Nuclear Information System (INIS)

    Basra, Gurpreet; Kirwan, Barry

    1998-01-01

    Accidents such as Piper Alpha have increased concern about the effects of human errors in complex systems. Such accidents can in theory be predicted and prevented by risk assessment, and in particular human reliability assessment (HRA), but HRA ideally requires qualitative and quantitative human error data. A research initiative at the University of Birmingham led to the development of CORE-DATA, a Computerised Human Error Data Base. This system currently contains a reasonably large number of human error data points, collected from a variety of mainly nuclear-power related sources. This article outlines a recent offshore data collection study, concerned with collecting lifeboat evacuation data. Data collection methods are outlined and a selection of human error probabilities generated as a result of the study are provided. These data give insights into the type of errors and human failure rates that could be utilised to support offshore risk analyses

  10. The collection and interpretation of domestic accident data. A discussion on some aspects of the British consumer safety system.

    Science.gov (United States)

    Wilson, J R

    1979-06-01

    This article discusses the workings of a Consumer Safety System and identifies the problems of collecting and using domestic accident data. In the light of proposed changes in the civil law on product liability, particular consideration is given to the difficulties of providing evidence of consumer product involvement in domestic accidents. The paper is based on one read at the CBI Conference "Product Liability in Perspective", held at the Hilton Hotel, London, 30-31 March, 1977. The views expressed are those of the Institute for Consumer Ergonomics and are not necessarily those of the Department of Prices and Consumer Protection.

  11. QuaDoSta - a freely configurable system which facilitates multi-centric data collection for healthcare and medical research

    Directory of Open Access Journals (Sweden)

    Albrecht, Ulrike

    2007-07-01

    Full Text Available This article describes QuaDoSta (quality assurance, documentation and statistics, a flexible documentation system as well as a data collection and networking platform for medical facilities. The user can freely define the required documentation masks which are easily expandable and can be adapted to individual requirements without the need for additional programming. To avoid duplication, data transfer interfaces can be configured flexibly to external sources such as patient management systems used in surgeries or hospital information systems. The projects EvaMed (Evaluation Anthroposophical Medicine and the Network Oncology are two scientific research projects which have been successfully established as nationally active networks on the basis of QuaDoSta. The EvaMed-Network serves as a modern pharmacovigilance project for the documentation of adverse drug events. All prescription data are electronically recorded to assess the relative risk of drugs. The Network Oncology was set up as a documentation system in four hospitals and seven specialist oncology practices where a complete record of all oncological therapies is being carried out to uniform standards on the basis of the ‘basic documentation for tumour patients’ (BDT developed by the German Cancer Society. The QuaDoSta solution system made it possible to cater for the specific requirements of the presented projects. The following features of the system proved to be highly advantageous: flexible setup of catalogues and user friendly customisation and extensions, complete dissociation of system setup and documentation content, multi-centre networkability, and configurable data transfer interfaces.

  12. Using the LANDSAT data collection system for field geophysics: Operations in the British Virgin Islands. [relaying borehole tiltmeter data

    Science.gov (United States)

    Webster, W. J., Jr.; Allen, W. K.; Gilbert, E. L.; Painter, J. E. (Principal Investigator)

    1980-01-01

    This particular application was to vertical geodesy by tide gauge and tiltmeter on a small desert island in the British Virgin Islands. The performance of the LANDSAT system under potentially marginal circumstances was found to be excellent.

  13. New insights into health financing: First results of the international data collection under the System of Health Accounts 2011 framework.

    Science.gov (United States)

    Mueller, Michael; Morgan, David

    2017-07-01

    International comparisons of health spending and financing are most frequently carried out using datasets of international organisations based on the System of Health Accounts (SHA). This accounting framework has recently been updated and 2016 saw the first international data collection under the new SHA 2011 guidelines. In addition to reaching better comparability of health spending figures and greater country coverage, the updated framework has seen changes in the dimension of health financing leading to important consequences when analysing health financing data. This article presents the first results of health spending and financing data collected under this new framework and highlights the areas where SHA 2011 has become a more useful tool for policy analysis, by complementing data on expenditure of health financing schemes with information about their revenue streams. It describes the major conceptual changes in the scope of health financing and highlights why comprehensive analyses based on SHA 2011 can provide for a more complete description and comparison of health financing across countries, facilitate a more meaningful discussion of fiscal sustainability of health spending by also analysing the revenues of compulsory public schemes and help to clarify the role of governments in financing health care - which is generally much bigger than previously documented. Copyright © 2017 Elsevier B.V. All rights reserved.

  14. Automatic Battery Swap System for Home Robots

    Directory of Open Access Journals (Sweden)

    Juan Wu

    2012-12-01

    Full Text Available This paper presents the design and implementation of an automatic battery swap system for the prolonged activities of home robots. A battery swap station is proposed to implement battery off-line recharging and on-line exchanging functions. It consists of a loading and unloading mechanism, a shifting mechanism, a locking device and a shell. The home robot is a palm-sized wheeled robot with an onboard camera and a removable battery case in the front. It communicates with the battery swap station wirelessly through ZigBee. The influences of battery case deflection and robot docking deflection on the battery swap operations have been investigated. The experimental results show that it takes an average time of 84.2s to complete the battery swap operations. The home robot does not have to wait several hours for the batteries to be fully charged. The proposed battery swap system is proved to be efficient in home robot applications that need the robots to work continuously over a long period.

  15. Evaluated data collections from ENSDF

    International Nuclear Information System (INIS)

    Ewbank, W.B.

    1979-01-01

    For several years the Nuclear Data Project has been maintaining an Evaluated Nuclear Structure Data File (ENSDF), which is designed to include critically evaluated values for most nuclear spectroscopic quantities. The information in ENSDF is the same as in the Nuclear Data Sheets, which illustrates two particular output formats (drawings and tables). Spectroscopic information for nuclei with A < 45 is put into ENSDF from the evaluations of Aizenberg-Selove and of Endt and van der Leun. An international network was organized to provide regular revisions of the data file. Computer facilities were developed to retrieve collections of evaluated data for special calculations or detailed examination

  16. AUTOMATIC EXTRACTION OF ROAD MARKINGS FROM MOBILE LASER SCANNING DATA

    Directory of Open Access Journals (Sweden)

    H. Ma

    2017-09-01

    Full Text Available Road markings as critical feature in high-defination maps, which are Advanced Driver Assistance System (ADAS and self-driving technology required, have important functions in providing guidance and information to moving cars. Mobile laser scanning (MLS system is an effective way to obtain the 3D information of the road surface, including road markings, at highway speeds and at less than traditional survey costs. This paper presents a novel method to automatically extract road markings from MLS point clouds. Ground points are first filtered from raw input point clouds using neighborhood elevation consistency method. The basic assumption of the method is that the road surface is smooth. Points with small elevation-difference between neighborhood are considered to be ground points. Then ground points are partitioned into a set of profiles according to trajectory data. The intensity histogram of points in each profile is generated to find intensity jumps in certain threshold which inversely to laser distance. The separated points are used as seed points to region grow based on intensity so as to obtain road mark of integrity. We use the point cloud template-matching method to refine the road marking candidates via removing the noise clusters with low correlation coefficient. During experiment with a MLS point set of about 2 kilometres in a city center, our method provides a promising solution to the road markings extraction from MLS data.

  17. Automatic Extraction of Road Markings from Mobile Laser Scanning Data

    Science.gov (United States)

    Ma, H.; Pei, Z.; Wei, Z.; Zhong, R.

    2017-09-01

    Road markings as critical feature in high-defination maps, which are Advanced Driver Assistance System (ADAS) and self-driving technology required, have important functions in providing guidance and information to moving cars. Mobile laser scanning (MLS) system is an effective way to obtain the 3D information of the road surface, including road markings, at highway speeds and at less than traditional survey costs. This paper presents a novel method to automatically extract road markings from MLS point clouds. Ground points are first filtered from raw input point clouds using neighborhood elevation consistency method. The basic assumption of the method is that the road surface is smooth. Points with small elevation-difference between neighborhood are considered to be ground points. Then ground points are partitioned into a set of profiles according to trajectory data. The intensity histogram of points in each profile is generated to find intensity jumps in certain threshold which inversely to laser distance. The separated points are used as seed points to region grow based on intensity so as to obtain road mark of integrity. We use the point cloud template-matching method to refine the road marking candidates via removing the noise clusters with low correlation coefficient. During experiment with a MLS point set of about 2 kilometres in a city center, our method provides a promising solution to the road markings extraction from MLS data.

  18. Automatic Compound Annotation from Mass Spectrometry Data Using MAGMa.

    NARCIS (Netherlands)

    Ridder, L.O.; Hooft, van der J.J.J.; Verhoeven, S.

    2014-01-01

    The MAGMa software for automatic annotation of mass spectrometry based fragmentation data was applied to 16 MS/MS datasets of the CASMI 2013 contest. Eight solutions were submitted in category 1 (molecular formula assignments) and twelve in category 2 (molecular structure assignment). The MS/MS

  19. TFTR data management system

    International Nuclear Information System (INIS)

    Randerson, L.; Chu, J.; Ludescher, C.; Malsbury, J.; Stark, W.

    1986-01-01

    Developments in the tokamak fusion test reactor (TFTR) data-management system supporting data acquisition and off-line physics data reduction are described. Data from monitor points, timing channels, transient recorder channels, and other devices are acquired and stored for use by on-line tasks. Files are transferred off line automatically. A configuration utility determines data acquired and files transferred. An event system driven by file arrival activates off-line reduction processes. A post-run process transfers files not shipped during runs. Files are archived to tape and are retrievable by digraph and shot number. Automatic skimming based on most recent access, file type, shot numbers, and user-set protections maintains the files required for post-run data reduction

  20. An automatic system for elaboration of chip breaking diagrams

    DEFF Research Database (Denmark)

    Andreasen, Jan Lasson; De Chiffre, Leonardo

    1998-01-01

    A laboratory system for fully automatic elaboration of chip breaking diagrams has been developed and tested. The system is based on automatic chip breaking detection by frequency analysis of cutting forces in connection with programming of a CNC-lathe to scan different feeds, speeds and cutting...

  1. Reactor protection system with automatic self-testing and diagnostic

    International Nuclear Information System (INIS)

    Gaubatz, D.C.

    1996-01-01

    A reactor protection system is disclosed having four divisions, with quad redundant sensors for each scram parameter providing input to four independent microprocessor-based electronic chassis. Each electronic chassis acquires the scram parameter data from its own sensor, digitizes the information, and then transmits the sensor reading to the other three electronic chassis via optical fibers. To increase system availability and reduce false scrams, the reactor protection system employs two levels of voting on a need for reactor scram. The electronic chassis perform software divisional data processing, vote 2/3 with spare based upon information from all four sensors, and send the divisional scram signals to the hardware logic panel, which performs a 2/4 division vote on whether or not to initiate a reactor scram. Each chassis makes a divisional scram decision based on data from all sensors. Automatic detection and discrimination against failed sensors allows the reactor protection system to automatically enter a known state when sensor failures occur. Cross communication of sensor readings allows comparison of four theoretically ''identical'' values. This permits identification of sensor errors such as drift or malfunction. A diagnostic request for service is issued for errant sensor data. Automated self test and diagnostic monitoring, sensor input through output relay logic, virtually eliminate the need for manual surveillance testing. This provides an ability for each division to cross-check all divisions and to sense failures of the hardware logic. 16 figs

  2. Developing a broadband automatic speech recognition system for Afrikaans

    CSIR Research Space (South Africa)

    De Wet, Febe

    2011-08-01

    Full Text Available baseline transcription for the news data. The match between a baseline transcription and its corre- sponding audio can be evaluated automatically using an ASR system in forced alignment mode. Only those bulletins for which a bad match is indicated... Component Index for data [3]. occurrence of Afrikaans words3. Other text corpora that are currently under construction in- clude daily downloads of the scripts of news bulletins that are read on an Afrikaans radio station as well as transcripts of par...

  3. Qualitative data collection with children.

    Science.gov (United States)

    Spratling, Regena; Coke, Sallie; Minick, Ptlene

    2012-02-01

    Qualitative researchers have clear methods to guide them in data collection with adult participants, but little is known about effective interview techniques with children. The findings from this methodological study on qualitative interviews with children indicate that children are able to articulate their experiences in interviews. Copyright © 2012 Elsevier Inc. All rights reserved.

  4. Design of a real-time tax-data monitoring intelligent card system

    Science.gov (United States)

    Gu, Yajun; Bi, Guotang; Chen, Liwei; Wang, Zhiyuan

    2009-07-01

    To solve the current problem of low efficiency of domestic Oil Station's information management, Oil Station's realtime tax data monitoring system has been developed to automatically access tax data of Oil pumping machines, realizing Oil-pumping machines' real-time automatic data collection, displaying and saving. The monitoring system uses the noncontact intelligent card or network to directly collect data which can not be artificially modified and so seals the loopholes and improves the tax collection's automatic level. It can perform real-time collection and management of the Oil Station information, and find the problem promptly, achieves the automatic management for the entire process covering Oil sales accounting and reporting. It can also perform remote query to the Oil Station's operation data. This system has broad application future and economic value.

  5. Scheduling algorithms for automatic control systems for technological processes

    Science.gov (United States)

    Chernigovskiy, A. S.; Tsarev, R. Yu; Kapulin, D. V.

    2017-01-01

    Wide use of automatic process control systems and the usage of high-performance systems containing a number of computers (processors) give opportunities for creation of high-quality and fast production that increases competitiveness of an enterprise. Exact and fast calculations, control computation, and processing of the big data arrays - all of this requires the high level of productivity and, at the same time, minimum time of data handling and result receiving. In order to reach the best time, it is necessary not only to use computing resources optimally, but also to design and develop the software so that time gain will be maximal. For this purpose task (jobs or operations), scheduling techniques for the multi-machine/multiprocessor systems are applied. Some of basic task scheduling methods for the multi-machine process control systems are considered in this paper, their advantages and disadvantages come to light, and also some usage considerations, in case of the software for automatic process control systems developing, are made.

  6. Research on Automatic Positioning System of Ultrasonic Testing of Wind Turbine Blade Flaws

    Science.gov (United States)

    Liu, Q. X.; Wang, Z. H.; Long, S. G.; Cai, M.; Cai, M.; Wang, X.; Chen, X. Y.; Bu, J. L.

    2017-11-01

    Ultrasonic testing technology has been used essentially in non-destructive testing of wind turbine blades. However, it is fact that the ultrasonic flaw detection method has inefficiently employed in recent years. This is because the testing result will illustrate a small deviation due to the artificial, environmental and technical factors. Therefore, it is an urgent technical demand for engineers to test the various flaws efficiently and quickly. An automatic positioning system has been designed in this paper to record the moving coordinates and the target distance in real time. Simultaneously, it could launch and acquire the sonic wave automatically. The ADNS-3080 optoelectronic chip is manufactured by Agilent Technologies Inc, which is also utilized in the system. With the combination of the chip, the power conversion module and the USB transmission module, the collected data can be transmitted from the upper monitor to the hardware that could process and control the data through software programming. An experiment has been designed to prove the reliability of automotive positioning system. The result has been validated by comparing the result collected form LABVIEW and actual plots on Perspex plane, it concludes that the system possesses high accuracy and magnificent meanings in practical engineering.

  7. Meteorological observatory for Antarctic data collection

    International Nuclear Information System (INIS)

    Grigioni, P.; De Silvestri, L.

    1996-01-01

    In the last years, a great number of automatic weather stations was installed in Antarctica, with the aim to examine closely the weather and climate of this region and to improve the coverage of measuring points on the Antarctic surface. In 1987 the Italian Antarctic Project started to set up a meteorological network, in an area not completely covered by other countries. Some of the activities performed by the meteorological observatory, concerning technical functions such as maintenance of the AWS's and the execution of radio soundings, or relating to scientific purposes such as validation and elaboration of collected data, are exposed. Finally, some climatological considerations on the thermal behaviour of the Antarctic troposphere such as 'coreless winter', and on the wind field, including katabatic flows in North Victoria Land are described

  8. Fully automatic AI-based leak detection system

    Energy Technology Data Exchange (ETDEWEB)

    Tylman, Wojciech; Kolczynski, Jakub [Dept. of Microelectronics and Computer Science, Technical University of Lodz in Poland, ul. Wolczanska 221/223, Lodz (Poland); Anders, George J. [Kinectrics Inc., 800 Kipling Ave., Toronto, Ontario M8Z 6C4 (Canada)

    2010-09-15

    This paper presents a fully automatic system intended to detect leaks of dielectric fluid in underground high-pressure, fluid-filled (HPFF) cables. The system combines a number of artificial intelligence (AI) and data processing techniques to achieve high detection capabilities for various rates of leaks, including leaks as small as 15 l per hour. The system achieves this level of precision mainly thanks to a novel auto-tuning procedure, enabling learning of the Bayesian network - the decision-making component of the system - using simulated leaks of various rates. Significant new developments extending the capabilities of the original leak detection system described in and form the basis of this paper. Tests conducted on the real-life HPFF cable system in New York City are also discussed. (author)

  9. Data collection modules for the PHENIX experiment

    International Nuclear Information System (INIS)

    Chi, C.Y.; Cole, B.; Nagle, J.L.; Sippach, W.; Zajc, W.A.

    1998-01-01

    The data acquisition (DAQ) system for the PHENIX experiment is designed as a pipeline system with simultaneous triggering and readout. The maximum average level-1 (LVL1) trigger rate is 25 KHz. The DAQ system consists of Front-End Modules (FEM's), a level-1 (LVL1) trigger, data collection modules (DCM's) timing systems, slow controllers and an event builder (EVB). The data collection modules have the responsibility of collecting uncompressed LVL1 trigger event fragments from the FEM's. The DCM's provide buffering for up to five LVL1 events. The DCM's also perform zero suppression, error checking, data reformatting and outputting data to the event builder. In addition to the FEM data, the DCM's also receive primitives from LVL1 trigger system. These primitives are used for alignment checking on the FEM data packet. Additional trigger primitives can also be generated together with the FEM data. The DCM is hosted in a VME crate. VME is used as a means for maintenance and slow control. Data collection within the crate is done through a private data-way

  10. Students Collecting Real time Data

    Science.gov (United States)

    Miller, P.

    2006-05-01

    Students Collecting Real-Time Data The Hawaiian Islands Humpback Whale National Marine Sanctuary has created opportunities for middle and high school students to become Student Researchers and to be involved in real-time marine data collection. It is important that we expose students to different fields of science and encourage them to enter scientific fields of study. The Humpback Whale Sanctuary has an education visitor center in Kihei, Maui. Located right on the beach, the site has become a living classroom facility. There is a traditional Hawaiian fishpond fronting the property. The fishpond wall is being restored, using traditional methods. The site has the incredible opportunity of incorporating Hawaiian cultural practices with scientific studies. The Sanctuary offers opportunities for students to get involved in monitoring and data collection studies. Invasive Seaweed Study: Students are collecting data on invasive seaweed for the University of Hawaii. They pull a large net through the shallow waters. Seaweed is sorted, identified and weighed. The invasive seaweeds are removed. The data is recorded and sent to UH. Remote controlled monitoring boats: The sanctuary has 6 boogie board sized remote controlled boats used to monitor reefs. Boats have a camera with lights on the underside. The boats have water quality monitoring devices and GPS units. The video from the underwater camera is transmitted via a wireless transmission. Students are able to monitor the fish, limu and invertebrate populations on the reef and collect water quality data via television monitors or computers. The boat can also pull a small plankton tow net. Data is being compiled into data bases. Artificial Reef Modules: The Sanctuary has a scientific permit from the state to build and deploy artificial reef modules. High school students are designing and building modules. These are deployed out in the Fishpond fronting the Sanctuary site and students are monitoring them on a weekly basis

  11. User Metrics in NASA Earth Science Data Systems

    Science.gov (United States)

    Lynnes, Chris

    2018-01-01

    This presentation the collection and use of user metrics in NASA's Earth Science data systems. A variety of collection methods is discussed, with particular emphasis given to the American Customer Satisfaction Index (ASCI). User sentiment on potential use of cloud computing is presented, with generally positive responses. The presentation also discusses various forms of automatically collected metrics, including an example of the relative usage of different functions within the Giovanni analysis system.

  12. Optical Automatic Car Identification (OACI) : Volume 1. Advanced System Specification.

    Science.gov (United States)

    1978-12-01

    A performance specification is provided in this report for an Optical Automatic Car Identification (OACI) scanner system which features 6% improved readability over existing industry scanner systems. It also includes the analysis and rationale which ...

  13. 47 CFR 25.281 - Automatic Transmitter Identification System (ATIS).

    Science.gov (United States)

    2010-10-01

    ... 47 Telecommunication 2 2010-10-01 2010-10-01 false Automatic Transmitter Identification System (ATIS). 25.281 Section 25.281 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Technical Operations § 25.281 Automatic Transmitter...

  14. Automatic early warning systems for the environment

    Directory of Open Access Journals (Sweden)

    Lesjak Martin

    2003-01-01

    Full Text Available Computerized, continuous monitoring environmental early warning systems are complex networks that merge measurements with the information technology. Accuracy, consistency, reliability and data quality are their most important features. Several effects may disturb their characteristics: hostile environment, unreliable communications, poor quality of equipment nonqualified users or service personnel. According to our experiences, a number of measures should be taken to enhance system performances and to maintain them at the desired level. In the paper, we are presenting an analysis of system requirements, possible disturbances and corrective measures that give the main directives for the design, construction and exploitation of the environmental early warning systems. Procedures which ensure data integrity and quality are mentioned. Finally, the contemporary system approach based on the LAN/WAN network topology with Intranet/Internet software is proposed, together with case descriptions of two already operating systems, based on computer-network principle.

  15. Automatic test system of the photomultipliers

    International Nuclear Information System (INIS)

    Shiino, Kazuo; Kono, Koji; Ishii, Takanobu; Kasai, Seiji; Yamada, Sakue; Kitamura, Shoichi.

    1990-03-01

    A test system of R580 photomultipliers (PMTs) was constructed for the ZEUS experiment HERA. In this report, we will describe the general feature of the test system, each component of the setup, the procedure of the measurements, the data analyses and the results of the first 800 PMT measurements. (author)

  16. Sodium component reliability data collection at CREDO

    International Nuclear Information System (INIS)

    Bott, T.F.; Haas, P.M.; Manning, J.J.

    1979-01-01

    The Centralized Reliability Data Organization (CREDO) has been established at Oak Ridge National Laboratory (ORNL) by the Department of Energy to provide a national center for collection, evaluation and dissemination of reliability data for advanced reactors. While the system is being developed and continuous data collection at the two U.S. reactor sites (EBR-II and FFTF) is being established, data on advanced reactor components which have been in use at U.S. test loops and experimental reactors have been collected and analyzed. Engineering, operating and event data on sodium valves, pumps, flow meters, rupture discs, heat exchangers and cold traps have been collected from more than a dozen sites. The results of analyses of the data performed to date are presented

  17. Automatically exposing OpenLifeData via SADI semantic Web Services.

    Science.gov (United States)

    González, Alejandro Rodríguez; Callahan, Alison; Cruz-Toledo, José; Garcia, Adrian; Egaña Aranguren, Mikel; Dumontier, Michel; Wilkinson, Mark D

    2014-01-01

    Two distinct trends are emerging with respect to how data is shared, collected, and analyzed within the bioinformatics community. First, Linked Data, exposed as SPARQL endpoints, promises to make data easier to collect and integrate by moving towards the harmonization of data syntax, descriptive vocabularies, and identifiers, as well as providing a standardized mechanism for data access. Second, Web Services, often linked together into workflows, normalize data access and create transparent, reproducible scientific methodologies that can, in principle, be re-used and customized to suit new scientific questions. Constructing queries that traverse semantically-rich Linked Data requires substantial expertise, yet traditional RESTful or SOAP Web Services cannot adequately describe the content of a SPARQL endpoint. We propose that content-driven Semantic Web Services can enable facile discovery of Linked Data, independent of their location. We use a well-curated Linked Dataset - OpenLifeData - and utilize its descriptive metadata to automatically configure a series of more than 22,000 Semantic Web Services that expose all of its content via the SADI set of design principles. The OpenLifeData SADI services are discoverable via queries to the SHARE registry and easy to integrate into new or existing bioinformatics workflows and analytical pipelines. We demonstrate the utility of this system through comparison of Web Service-mediated data access with traditional SPARQL, and note that this approach not only simplifies data retrieval, but simultaneously provides protection against resource-intensive queries. We show, through a variety of different clients and examples of varying complexity, that data from the myriad OpenLifeData can be recovered without any need for prior-knowledge of the content or structure of the SPARQL endpoints. We also demonstrate that, via clients such as SHARE, the complexity of federated SPARQL queries is dramatically reduced.

  18. Toward Global Comparability of Sexual Orientation Data in Official Statistics: A Conceptual Framework of Sexual Orientation for Health Data Collection in New Zealand's Official Statistics System

    Science.gov (United States)

    Gray, Alistair; Veale, Jaimie F.; Binson, Diane; Sell, Randell L.

    2013-01-01

    Objective. Effectively addressing health disparities experienced by sexual minority populations requires high-quality official data on sexual orientation. We developed a conceptual framework of sexual orientation to improve the quality of sexual orientation data in New Zealand's Official Statistics System. Methods. We reviewed conceptual and methodological literature, culminating in a draft framework. To improve the framework, we held focus groups and key-informant interviews with sexual minority stakeholders and producers and consumers of official statistics. An advisory board of experts provided additional guidance. Results. The framework proposes working definitions of the sexual orientation topic and measurement concepts, describes dimensions of the measurement concepts, discusses variables framing the measurement concepts, and outlines conceptual grey areas. Conclusion. The framework proposes standard definitions and concepts for the collection of official sexual orientation data in New Zealand. It presents a model for producers of official statistics in other countries, who wish to improve the quality of health data on their citizens. PMID:23840231

  19. Automatic control system in the reactor peggy

    International Nuclear Information System (INIS)

    Bertrand, J.; Mourchon, R.; Da Costa, D.; Desandre-Navarre, Ch.

    1967-01-01

    The equipment makes it possible for the reactor to attain a given power automatically and for the power to be maintained around this level. The principle of its operation consists in the changing from one power to another, at constant period, by means of a programmer transforming a power-step request into a voltage variation which is linear with time and which represents the logarithm of the required power. The real power is compared continuously with the required power. Stabilization occurs automatically as soon as the difference between the reactor power and the required power diminishes to a few per cent. (authors) [fr

  20. Usage of aids monitoring in automatic braking systems of modern cars

    Directory of Open Access Journals (Sweden)

    Dembitskyi V.

    2016-08-01

    Full Text Available Increased safety can be carried out at the expense the installation on vehicles of automatic braking systems, that monitor the traffic situation and the actions of the driver. In this paper considered the advantages and disadvantages of automatic braking systems, were analyzed modern tracking tools that are used in automatic braking systems. Based on the statistical data on accidents, are set the main dangers, that the automatic braking system will be reduced. In order to ensure the accuracy of information conducted research for determination of optimal combination of different sensors that provide an adequate perception of road conditions. The tracking system should be equipped with a combination of sensors, which in the case of detection of an obstacle or dangers of signal is transmitted to the information processing system and decision making. Information from the monitoring system should include data for the identification of the object, its condition, the speed.

  1. NASA Scientific Data Purchase Project: From Collection to User

    Science.gov (United States)

    Nicholson, Lamar; Policelli, Fritz; Fletcher, Rose

    2002-01-01

    NASA's Scientific Data Purchase (SDP) project is currently a $70 million operation managed by the Earth Science Applications Directorate at Stennis Space Center. The SDP project was developed in 1997 to purchase scientific data from commercial sources for distribution to NASA Earth science researchers. Our current data holdings include 8TB of remote sensing imagery consisting of 18 products from 4 companies. Our anticipated data volume is 60 TB by 2004, and we will be receiving new data products from several additional companies. Our current system capacity is 24 TB, expandable to 89 TB. Operations include tasking of new data collections, archive ordering, shipment verification, data validation, distribution, metrics, finances, customer feedback, and technical support. The program has been included in the Stennis Space Center Commercial Remote Sensing ISO 9001 registration since its inception. Our operational system includes automatic quality control checks on data received (with MatLab analysis); internally developed, custom Web-based interfaces that tie into commercial-off-the-shelf software; and an integrated relational database that links and tracks all data through operations. We've distributed nearly 1500 datasets, and almost 18,000 data files have been downloaded from our public web site; on a 10-point scale, our customer satisfaction index is 8.32 at a 23% response level. More information about the SDP is available on our Web site.

  2. Simple Approaches to Improve the Automatic Inventory of ZEBRA Crossing from Mls Data

    Science.gov (United States)

    Arias, P.; Riveiro, B.; Soilán, M.; Díaz-Vilariño, L.; Martínez-Sánchez, J.

    2015-08-01

    The city management is increasingly supported by information technologies, leading to paradigms such as smart cities, where decision-makers, companies and citizens are continuously interconnected. 3D modelling turns of great relevance when the city has to be managed making use of geospatial databases or Geographic Information Systems. On the other hand, laser scanning technology has experienced a significant growth in the last years, and particularly, terrestrial mobile laser scanning platforms are being more and more used with inventory purposes in both cities and road environments. Consequently, large datasets are available to produce the geometric basis for the city model; however, this data is not directly exploitable by management systems constraining the implementation of the technology for such applications. This paper presents a new algorithm for the automatic detection of zebra crossing. The algorithm is divided in three main steps: road segmentation (based on a PCA analysis of the points contained in each cycle of collected by a mobile laser system), rasterization (conversion of the point cloud to a raster image coloured as a function of intensity data), and zebra crossing detection (using the Hough Transform and logical constrains for line classification). After evaluating different datasets collected in three cities located in Northwest Spain (comprising 25 strips with 30 visible zebra crossings) a completeness of 83% was achieved.

  3. Automatic shading effects on the energetic performance of building systems; Efeito do sombreamento automatico no desempenho de sistemas prediais

    Energy Technology Data Exchange (ETDEWEB)

    Prado, Racine Tadeu Araujo

    1997-12-31

    This thesis develops a theoretic-experimental study dealing with the effects of an automatic shading device on the energetic performance of a dimmable lighting system and a cooling equipment. Some equations related to fenestration optical and thermal properties are rebuilt, while some others are created, under a theoretical approach. In order to collect field data, the energy demand-and other variables - was measured in two distinct stories, with the same fenestration features, of the Test Tower. New data was gathered after adding an automatic shading device to the window of one story. The comparison of the collected data allows the energetic performance evaluation of the shading device. (author) 136 refs., 55 figs., 6 tabs.

  4. Automatic shading effects on the energetic performance of building systems; Efeito do sombreamento automatico no desempenho de sistemas prediais

    Energy Technology Data Exchange (ETDEWEB)

    Prado, Racine Tadeu Araujo

    1996-12-31

    This thesis develops a theoretic-experimental study dealing with the effects of an automatic shading device on the energetic performance of a dimmable lighting system and a cooling equipment. Some equations related to fenestration optical and thermal properties are rebuilt, while some others are created, under a theoretical approach. In order to collect field data, the energy demand-and other variables - was measured in two distinct stories, with the same fenestration features, of the Test Tower. New data was gathered after adding an automatic shading device to the window of one story. The comparison of the collected data allows the energetic performance evaluation of the shading device. (author) 136 refs., 55 figs., 6 tabs.

  5. Usefulness of a centralized system of data collection for the development of an international multicentre registry of spondyloarthritis

    Science.gov (United States)

    Schiotis, Ruxandra; Font, Pilar; Zarco, Pedro; Almodovar, Raquel; Gratacós, Jordi; Mulero, Juan; Juanola, Xavier; Montilla, Carlos; Moreno, Estefanía; Ariza Ariza, Rafael; Collantes-Estevez, Eduardo

    2011-01-01

    Objective. To present the usefulness of a centralized system of data collection for the development of an international multicentre registry of SpA. Method. The originality of this registry consists in the creation of a virtual network of researchers in a computerized Internet database. From its conception, the registry was meant to be a dynamic acquiring system. Results. REGISPONSER has two developing phases (Conception and Universalization) and gathers several evolving secondary projects (REGISPONSER-EARLY, REGISPONSER-AS, ESPERANZA and RESPONDIA). Each sub-project answered the necessity of having more specific and complete data of the patients even from the onset of the disease so, in the end, obtaining a well-defined picture of SpAs spectrum in the Spanish population. Conclusion. REGISPONSER is the first dynamic SpA database composed of cohorts with a significant number of patients distributed by specific diagnosis, which provides basic specific information of the sub-cohorts useful for patients’ evaluation in rheumatology ambulatory consulting. PMID:20823095

  6. Computer control in nondestructive testing illustrated by an automatic ultrasonic tube inspection system

    International Nuclear Information System (INIS)

    Gundtoft, H.E.; Nielsen, N.

    1976-06-01

    In Risoe's automatic tube inspection system, data (more than half a million per tube) from ultrasonic dimension measurements and defect inspections are fed into a computer that simultaneously calculates and evaluates the results. (author)

  7. Automatic protein structure solution from weak X-ray data

    Science.gov (United States)

    Skubák, Pavol; Pannu, Navraj S.

    2013-11-01

    Determining new protein structures from X-ray diffraction data at low resolution or with a weak anomalous signal is a difficult and often an impossible task. Here we propose a multivariate algorithm that simultaneously combines the structure determination steps. In tests on over 140 real data sets from the protein data bank, we show that this combined approach can automatically build models where current algorithms fail, including an anisotropically diffracting 3.88 Å RNA polymerase II data set. The method seamlessly automates the process, is ideal for non-specialists and provides a mathematical framework for successfully combining various sources of information in image processing.

  8. RAMS data collection under Arctic conditions

    International Nuclear Information System (INIS)

    Barabadi, Abbas; Tobias Gudmestad, Ove; Barabady, Javad

    2015-01-01

    Reliability, availability, maintainability and supportability analysis is an important step in the design and operation of production processes and technology. Historical data such as time between failures and time to repairs play an important role in such analysis. The data must reflect the conditions that equipment has experienced during its operating time. To have a precise understanding of the conditions experienced, all influence factors on the failure and repair processes of a production facility in Arctic environment need to be identified and collected in the database. However, there is a lack of attention to collect the effect of influence factors in the reliability, availability, maintainability and supportability database. Hence, the aim of this paper is to discuss the challenges of the available methods of data collection and suggest a methodology for data collection considering the effect of environmental conditions. Application of the methodology will make the historical RAMS data of a system more applicable and useful for the design and operation of the system in different types of operational environments. - Highlights: • The challenges related to use of the available RAMS data is discussed. • It is important to collect information about operational condition in RAMS data. • A methodology for RAMS data collection considering environment condition is suggested. • Information about influence factors will make the result of RAMS analysis more applicable

  9. Automatic computation of radioimmunoassay data. Insulin and C-peptide

    Energy Technology Data Exchange (ETDEWEB)

    Toyota, T; Kudo, M; Abe, K [Hirosaki Univ., Aomori (Japan). School of Medicine; Kawamata, F; Uehata, S

    1975-09-01

    Radioimmunoassay provided dose response curves which showed linearity by the use of logistic transformation (Rodbard). This transformation which was applicable to radioimmunoassay should be useful for the computer processing of insulin and C-peptide assay. In the present studies, standard curves were analysed by testing the fit of analytic functions to radioimmunoassay of insulin and C-peptides. A program for use in combination with the double antibody technique was made by Dr. Kawamata. This approach was evidenced to be useful in order to allow automatic computation of data derived from the double antibody assays of insulin and C-peptides. Automatic corrected calculations of radioimmunoassay data of insulin was found to be satisfactory.

  10. Automatic digital photo-book making system

    Science.gov (United States)

    Wang, Wiley; Teo, Patrick; Muzzolini, Russ

    2010-02-01

    The diversity of photo products has grown more than ever before. A group of photos are not only printed individually, but also can be arranged in specific order to tell a story, such as in a photo book, a calendar or a poster collage. Similar to making a traditional scrapbook, digital photo book tools allow the user to choose a book style/theme, layouts of pages, backgrounds and the way the pictures are arranged. This process is often time consuming to users, given the number of images and the choices of layout/background combinations. In this paper, we developed a system to automatically generate photo books with only a few initial selections required. The system utilizes time stamps, color indices, orientations and other image properties to best fit pictures into a final photo book. The common way of telling a story is to lay the pictures out in chronological order. If the pictures are proximate in time, they will coincide with each other and are often logically related. The pictures are naturally clustered along a time line. Breaks between clusters can be used as a guide to separate pages or spreads, thus, pictures that are logically related can stay close on the same page or spread. When people are making a photo book, it is helpful to start with chronologically grouped images, but time alone wont be enough to complete the process. Each page is limited by the number of layouts available. Many aesthetic rules also apply, such as, emphasis of preferred pictures, consistency of local image density throughout the whole book, matching a background to the content of the images, and the variety of adjacent page layouts. We developed an algorithm to group images onto pages under the constraints of aesthetic rules. We also apply content analysis based on the color and blurriness of each picture, to match backgrounds and to adjust page layouts. Some of our aesthetic rules are fixed and given by designers. Other aesthetic rules are statistic models trained by using

  11. Discrete Model Reference Adaptive Control System for Automatic Profiling Machine

    Directory of Open Access Journals (Sweden)

    Peng Song

    2012-01-01

    Full Text Available Automatic profiling machine is a movement system that has a high degree of parameter variation and high frequency of transient process, and it requires an accurate control in time. In this paper, the discrete model reference adaptive control system of automatic profiling machine is discussed. Firstly, the model of automatic profiling machine is presented according to the parameters of DC motor. Then the design of the discrete model reference adaptive control is proposed, and the control rules are proven. The results of simulation show that adaptive control system has favorable dynamic performances.

  12. Automatic latency equalization in VHDL-implemented complex pipelined systems

    Science.gov (United States)

    Zabołotny, Wojciech M.

    2016-09-01

    In the pipelined data processing systems it is very important to ensure that parallel paths delay data by the same number of clock cycles. If that condition is not met, the processing blocks receive data not properly aligned in time and produce incorrect results. Manual equalization of latencies is a tedious and error-prone work. This paper presents an automatic method of latency equalization in systems described in VHDL. The proposed method uses simulation to measure latencies and verify introduced correction. The solution is portable between different simulation and synthesis tools. The method does not increase the complexity of the synthesized design comparing to the solution based on manual latency adjustment. The example implementation of the proposed methodology together with a simple design demonstrating its use is available as an open source project under BSD license.

  13. Development of automatic flaw detection systems for magnetic particle examination

    International Nuclear Information System (INIS)

    Shirai, T.; Kimura, J.; Amako, T.

    1988-01-01

    Utilizing a video camera and an image processor, development was carried out on automatic flaw detection and discrimination techniques for the purpose of achieving automated magnetic particle examination. Following this, fluorescent wet magnetic particle examination systems for blade roots and rotor grooves of turbine rotors and the non-fluorescent dry magnetic particle examination system for butt welds, were developed. This paper describes these automatic magnetic particle examination (MT) systems and the functional test results

  14. An Intelligent Tool for Activity Data Collection

    Directory of Open Access Journals (Sweden)

    A. M. Jehad Sarkar

    2011-04-01

    Full Text Available Activity recognition systems using simple and ubiquitous sensors require a large variety of real-world sensor data for not only evaluating their performance but also training the systems for better functioning. However, a tremendous amount of effort is required to setup an environment for collecting such data. For example, expertise and resources are needed to design and install the sensors, controllers, network components, and middleware just to perform basic data collections. It is therefore desirable to have a data collection method that is inexpensive, flexible, user-friendly, and capable of providing large and diverse activity datasets. In this paper, we propose an intelligent activity data collection tool which has the ability to provide such datasets inexpensively without physically deploying the testbeds. It can be used as an inexpensive and alternative technique to collect human activity data. The tool provides a set of web interfaces to create a web-based activity data collection environment. It also provides a web-based experience sampling tool to take the user’s activity input. The tool generates an activity log using its activity knowledge and the user-given inputs. The activity knowledge is mined from the web. We have performed two experiments to validate the tool’s performance in producing reliable datasets.

  15. An intelligent tool for activity data collection.

    Science.gov (United States)

    Sarkar, A M Jehad

    2011-01-01

    Activity recognition systems using simple and ubiquitous sensors require a large variety of real-world sensor data for not only evaluating their performance but also training the systems for better functioning. However, a tremendous amount of effort is required to setup an environment for collecting such data. For example, expertise and resources are needed to design and install the sensors, controllers, network components, and middleware just to perform basic data collections. It is therefore desirable to have a data collection method that is inexpensive, flexible, user-friendly, and capable of providing large and diverse activity datasets. In this paper, we propose an intelligent activity data collection tool which has the ability to provide such datasets inexpensively without physically deploying the testbeds. It can be used as an inexpensive and alternative technique to collect human activity data. The tool provides a set of web interfaces to create a web-based activity data collection environment. It also provides a web-based experience sampling tool to take the user's activity input. The tool generates an activity log using its activity knowledge and the user-given inputs. The activity knowledge is mined from the web. We have performed two experiments to validate the tool's performance in producing reliable datasets.

  16. Automatic extraction of pavement markings on streets from point cloud data of mobile LiDAR

    International Nuclear Information System (INIS)

    Gao, Yang; Zhong, Ruofei; Liu, Xianlin; Tang, Tao; Wang, Liuzhao

    2017-01-01

    Pavement markings provide an important foundation as they help to keep roads users safe. Accurate and comprehensive information about pavement markings assists the road regulators and is useful in developing driverless technology. Mobile light detection and ranging (LiDAR) systems offer new opportunities to collect and process accurate pavement markings’ information. Mobile LiDAR systems can directly obtain the three-dimensional (3D) coordinates of an object, thus defining spatial data and the intensity of (3D) objects in a fast and efficient way. The RGB attribute information of data points can be obtained based on the panoramic camera in the system. In this paper, we present a novel method process to automatically extract pavement markings using multiple attribute information of the laser scanning point cloud from the mobile LiDAR data. This method process utilizes a differential grayscale of RGB color, laser pulse reflection intensity, and the differential intensity to identify and extract pavement markings. We utilized point cloud density to remove the noise and used morphological operations to eliminate the errors. In the application, we tested our method process on different sections of roads in Beijing, China, and Buffalo, NY, USA. The results indicated that both correctness ( p ) and completeness ( r ) were higher than 90%. The method process of this research can be applied to extract pavement markings from huge point cloud data produced by mobile LiDAR. (paper)

  17. Automatic extraction of pavement markings on streets from point cloud data of mobile LiDAR

    Science.gov (United States)

    Gao, Yang; Zhong, Ruofei; Tang, Tao; Wang, Liuzhao; Liu, Xianlin

    2017-08-01

    Pavement markings provide an important foundation as they help to keep roads users safe. Accurate and comprehensive information about pavement markings assists the road regulators and is useful in developing driverless technology. Mobile light detection and ranging (LiDAR) systems offer new opportunities to collect and process accurate pavement markings’ information. Mobile LiDAR systems can directly obtain the three-dimensional (3D) coordinates of an object, thus defining spatial data and the intensity of (3D) objects in a fast and efficient way. The RGB attribute information of data points can be obtained based on the panoramic camera in the system. In this paper, we present a novel method process to automatically extract pavement markings using multiple attribute information of the laser scanning point cloud from the mobile LiDAR data. This method process utilizes a differential grayscale of RGB color, laser pulse reflection intensity, and the differential intensity to identify and extract pavement markings. We utilized point cloud density to remove the noise and used morphological operations to eliminate the errors. In the application, we tested our method process on different sections of roads in Beijing, China, and Buffalo, NY, USA. The results indicated that both correctness (p) and completeness (r) were higher than 90%. The method process of this research can be applied to extract pavement markings from huge point cloud data produced by mobile LiDAR.

  18. Automatic diagnostic methods of nuclear reactor collected signals

    International Nuclear Information System (INIS)

    Lavison, P.

    1978-03-01

    This work is the first phase of an opwall study of diagnosis limited to problems of monitoring the operating state; this allows to show all what the pattern recognition methods bring at the processing level. The present problem is the research of the control operations. The analysis of the state of the reactor gives a decision which is compared with the history of the control operations, and if there is not correspondence, the state subjected to the analysis will be said 'abnormal''. The system subjected to the analysis is described and the problem to solve is defined. Then, one deals with the gaussian parametric approach and the methods to evaluate the error probability. After one deals with non parametric methods and an on-line detection has been tested experimentally. Finally a non linear transformation has been studied to reduce the error probability previously obtained. All the methods presented have been tested and compared to a quality index: the error probability [fr

  19. SABER-School Finance : Data Collection Instrument

    OpenAIRE

    World Bank

    2015-01-01

    The aim of the SABER-school finance initiative is to collect, analyze and disseminate comparable data about education finance systems across countries. SABER-school finance assesses education finance systems along six policy goals: (i) ensuring basic conditions for learning; (ii) monitoring learning conditions and outcomes; (iii) overseeing service delivery; (iv) budgeting with adequate an...

  20. Effectiveness of data collection and information transmission ...

    African Journals Online (AJOL)

    2012-12-24

    Dec 24, 2012 ... correctness of records, Health Management Information System records ... Conclusion: The health workers were not operating the DSN system in the State to optimal functionality. ... Key words: Data collection, disease notification, effectiveness, information transmission process .... facilities to support it.

  1. SABER-School Finance: Data Collection Instrument

    Science.gov (United States)

    King, Elizabeth; Patrinos, Harry; Rogers, Halsey

    2015-01-01

    The aim of the SABER-school finance initiative is to collect, analyze and disseminate comparable data about education finance systems across countries. SABER-school finance assesses education finance systems along six policy goals: (i) ensuring basic conditions for learning; (ii) monitoring learning conditions and outcomes; (iii) overseeing…

  2. An Automatic Control System for Conditioning 30 GHz Accelerating Structures

    CERN Document Server

    Dubrovskiy, A

    2008-01-01

    A software application programme has been developed to allow fast and automatic high-gradient conditioning of accelerating structures at 30 GHz in CTF3. The specificity of the application is the ability to control the high-power electron beam which produces the 30 GHz RF power used to condition the accelerating structures. The programme permits operation round the clock with minimum manpower requirements. In this paper the fast control system, machine control system, logging system, graphical user control interface and logging data visualization are described. An outline of the conditioning control system itself and of the feedback controlling peak power and pulse length is given. The software allows different types of conditioning strategies to be programmed

  3. Reconstruction of the sea surface elevation from the analysis of the data collected by a wave radar system

    Science.gov (United States)

    Ludeno, Giovanni; Soldovieri, Francesco; Serafino, Francesco; Lugni, Claudio; Fucile, Fabio; Bulian, Gabriele

    2016-04-01

    X-band radar system is able to provide information about direction and intensity of the sea surface currents and dominant waves in a range of few kilometers from the observation point (up to 3 nautical miles). This capability, together with their flexibility and low cost, makes these devices useful tools for the sea monitoring either coastal or off-shore area. The data collected from wave radar system can be analyzed by using the inversion strategy presented in [1,2] to obtain the estimation of the following sea parameters: peak wave direction; peak period; peak wavelength; significant wave height; sea surface current and bathymetry. The estimation of the significant wave height represents a limitation of the wave radar system because of the radar backscatter is not directly related to the sea surface elevation. In fact, in the last period, substantial research has been carried out to estimate significant wave height from radar images either with or without calibration using in-situ measurements. In this work, we will present two alternative approaches for the reconstruction of the sea surface elevation from wave radar images. In particular, the first approach is based on the basis of an approximated version of the modulation transfer function (MTF) tuned from a series of numerical simulation, following the line of[3]. The second approach is based on the inversion of radar images using a direct regularised least square technique. Assuming a linearised model for the tilt modulation, the sea elevation has been reconstructed as a least square fitting of the radar imaging data[4]. References [1]F. Serafino, C. Lugni, and F. Soldovieri, "A novel strategy for the surface current determination from marine X-band radar data," IEEE Geosci.Remote Sens. Lett., vol. 7, no. 2, pp. 231-235, Apr. 2010. [2]Ludeno, G., Brandini, C., Lugni, C., Arturi, D., Natale, A., Soldovieri, F., Serafino, F. (2014). Remocean System for the Detection of the Reflected Waves from the Costa

  4. Design of an automatic production monitoring system on job shop manufacturing

    Science.gov (United States)

    Prasetyo, Hoedi; Sugiarto, Yohanes; Rosyidi, Cucuk Nur

    2018-02-01

    Every production process requires monitoring system, so the desired efficiency and productivity can be monitored at any time. This system is also needed in the job shop type of manufacturing which is mainly influenced by the manufacturing lead time. Processing time is one of the factors that affect the manufacturing lead time. In a conventional company, the recording of processing time is done manually by the operator on a sheet of paper. This method is prone to errors. This paper aims to overcome this problem by creating a system which is able to record and monitor the processing time automatically. The solution is realized by utilizing electric current sensor, barcode, RFID, wireless network and windows-based application. An automatic monitoring device is attached to the production machine. It is equipped with a touch screen-LCD so that the operator can use it easily. Operator identity is recorded through RFID which is embedded in his ID card. The workpiece data are collected from the database by scanning the barcode listed on its monitoring sheet. A sensor is mounted on the machine to measure the actual machining time. The system's outputs are actual processing time and machine's capacity information. This system is connected wirelessly to a workshop planning application belongs to the firm. Test results indicated that all functions of the system can run properly. This system successfully enables supervisors, PPIC or higher level management staffs to monitor the processing time quickly with a better accuracy.

  5. AUTOMATIC CLASSIFICATION OF VARIABLE STARS IN CATALOGS WITH MISSING DATA

    International Nuclear Information System (INIS)

    Pichara, Karim; Protopapas, Pavlos

    2013-01-01

    We present an automatic classification method for astronomical catalogs with missing data. We use Bayesian networks and a probabilistic graphical model that allows us to perform inference to predict missing values given observed data and dependency relationships between variables. To learn a Bayesian network from incomplete data, we use an iterative algorithm that utilizes sampling methods and expectation maximization to estimate the distributions and probabilistic dependencies of variables from data with missing values. To test our model, we use three catalogs with missing data (SAGE, Two Micron All Sky Survey, and UBVI) and one complete catalog (MACHO). We examine how classification accuracy changes when information from missing data catalogs is included, how our method compares to traditional missing data approaches, and at what computational cost. Integrating these catalogs with missing data, we find that classification of variable objects improves by a few percent and by 15% for quasar detection while keeping the computational cost the same

  6. AUTOMATIC CLASSIFICATION OF VARIABLE STARS IN CATALOGS WITH MISSING DATA

    Energy Technology Data Exchange (ETDEWEB)

    Pichara, Karim [Computer Science Department, Pontificia Universidad Católica de Chile, Santiago (Chile); Protopapas, Pavlos [Institute for Applied Computational Science, Harvard University, Cambridge, MA (United States)

    2013-11-10

    We present an automatic classification method for astronomical catalogs with missing data. We use Bayesian networks and a probabilistic graphical model that allows us to perform inference to predict missing values given observed data and dependency relationships between variables. To learn a Bayesian network from incomplete data, we use an iterative algorithm that utilizes sampling methods and expectation maximization to estimate the distributions and probabilistic dependencies of variables from data with missing values. To test our model, we use three catalogs with missing data (SAGE, Two Micron All Sky Survey, and UBVI) and one complete catalog (MACHO). We examine how classification accuracy changes when information from missing data catalogs is included, how our method compares to traditional missing data approaches, and at what computational cost. Integrating these catalogs with missing data, we find that classification of variable objects improves by a few percent and by 15% for quasar detection while keeping the computational cost the same.

  7. Simultaneous real-time data collection methods

    Science.gov (United States)

    Klincsek, Thomas

    1992-01-01

    This paper describes the development of electronic test equipment which executes, supervises, and reports on various tests. This validation process uses computers to analyze test results and report conclusions. The test equipment consists of an electronics component and the data collection and reporting unit. The PC software, display screens, and real-time data-base are described. Pass-fail procedures and data replay are discussed. The OS2 operating system and Presentation Manager user interface system were used to create a highly interactive automated system. The system outputs are hardcopy printouts and MS DOS format files which may be used as input for other PC programs.

  8. [An automatic system controlled by microcontroller for carotid sinus perfusion].

    Science.gov (United States)

    Yi, X L; Wang, M Y; Fan, Z Z; He, R R

    2001-08-01

    To establish a new method for controlling automatically the carotid perfusion pressure. A cheap practical automatic perfusion unit based on AT89C2051 micro controller was designed. The unit, LDB-M perfusion pump and the carotid sinus of an animal constituted an automatic perfusion system. This system was able to provide ramp and stepwise updown perfusion pattern and has been used in the research of baroreflex. It can insure the precision and reproducibility of perfusion pressure curve, and improve the technical level in corresponding medical field.

  9. A Microcontroller-Based Automatic Transfer Switching System for a ...

    African Journals Online (AJOL)

    Michael

    2015-06-01

    Jun 1, 2015 ... Most industries still employ the manual method of power supply changeover, ... This paper presents a Microcontroller-Based Automatic Transfer Switching System ..... and currently has special research interest in Wireless.

  10. Fighter/Attack Automatic Collision Avoidance Systems Business Case

    National Research Council Canada - National Science Library

    Mapes, Peter B

    2006-01-01

    .... This study concludes that implementation of Automatic Collision Avoidance Systems (Auto-CAS) in F-16, F/A-18, F/A-22, and F-35 aircraft would save aircrew lives and preserve, and enhance combat capability.

  11. MAD data collection - current trends

    International Nuclear Information System (INIS)

    Dementieva, I.; Evans, G.; Joachimiak, A.; Sanishvili, R.; Walsh, M. A.

    1999-01-01

    The multi-wavelength anomalous diffraction, or MAD, method of determining protein structure is becoming routine in protein crystallography. An increase in the number of tuneable synchrotrons beamlines coupled with the widespread availability position-sensitive X-ray detectors based on charged-coupled devices and having fast readout raised MAD structure determination to a new and exciting level. Ultra-fast MAD data collection is now possible. Recognition of the value of selenium for phasing protein structures and improvement of methods for incorporating selenium into proteins in the form of selenomethionine have attracted greater interest in the MAD method. Recent developments in crystallographic software are complimenting the above advances, paving the way for rapid protein structure determination. An overview of a typical MAD experiment is described here, with emphasis on the rates and quality of data acquisition now achievable at beamlines developed at third-generation synchrotrons sources

  12. The Diagnostic System of A – 604 Automatic Transmission

    Directory of Open Access Journals (Sweden)

    Czaban Jaroslaw

    2014-09-01

    Full Text Available Automatic gearbox gains increasing popularity in Europe. Little interest in diagnosis of such type of transmission in Poland results from the fact of small share in the whole market of operated cars, so there is a lack of availability of special diagnostic devices. These factors cause issues of expensive repairs, often involving a replacement of subassembly to new or aftermarket one. To a small extent some prophylactic diagnostic tests are conducted, which can eliminate future gearbox system failures. In the paper, the proposition of diagnostic system of popular A - 604 gearbox was presented. The authors are seeking for the possibility of using such type of devices to functional elaboration of gearboxes after renovation. The built system pursues the drive of the researched object, connected with simulated load, where special controller, replacing the original one, is responsible for controlling gearbox operation. This way is used to evaluate the mechanic and hydraulic parts' state. Analysis of signal runs, registered during measurements lets conclude about operation correctness, where as comparison with stock data verifies the technical state of an automatic gearbox.

  13. Realization of automatic test system for induction motors

    Energy Technology Data Exchange (ETDEWEB)

    Adragna, R; Nuccio, S [Assessorato Regionale del Lavoro, Palermo (Italy). Centro Elaborazione Dati Palermo Univ. (Italy). Dip. di Ingegneria Elettrica

    1991-03-01

    In this paper, the general principles of the design of the hardware and software of an automatic test system, are outlined. With reference to the requirements of electric motor test laboratories, design principles, specific for such applications, are pointed out and an automatic test system for induction motors, developed at the Dipartimento di Ingegneria Elettrica of Palermo University, is described. The peculiarities of the hardware utilized and of the specific software developed are illustrated. Finally, some examples of applications are showed.

  14. Development of the automatic control rod operation system for JOYO. Verification of automatic control rod operation guide system

    International Nuclear Information System (INIS)

    Terakado, Tsuguo; Suzuki, Shinya; Kawai, Masashi; Aoki, Hiroshi; Ohkubo, Toshiyuki

    1999-10-01

    The automatic control rod operation system was developed to control the JOYO reactor power automatically in all operation modes(critical approach, cooling system heat up, power ascent, power descent), development began in 1989. Prior to applying the system, verification tests of the automatic control rod operation guide system was conducted during 32nd duty cycles of JOYO' from Dec. 1997 to Feb. 1998. The automatic control rod operation guide system consists of the control rod operation guide function and the plant operation guide function. The control rod operation guide function provides information on control rod movement and position, while the plant operation guide function provide guidance for plant operations corresponding to reactor power changes(power ascent or power descent). Control rod insertion or withdrawing are predicted by fuzzy algorithms. (J.P.N.)

  15. Characteristics and design improvement of AP1000 automatic depressurization system

    International Nuclear Information System (INIS)

    Jin Fei

    2012-01-01

    Automatic depressurization system, as a specialty of AP1000 Design, enhances capability of mitigating design basis accidents for plant. Advancement of the system is discussed by comparing with traditional PWR design and analyzing system functions, such as depressurizing and venting. System design improvement during China Project performance is also described. At the end, suggestions for the system in China Project are listed. (author)

  16. Adaptive intrusion data system

    International Nuclear Information System (INIS)

    Johnson, C.S.

    1976-01-01

    An Adaptive Intrusion Data System (AIDS) was developed to collect data from intrusion alarm sensors as part of an evaluation system to improve sensor performance. AIDS is a unique digital data compression, storage, and formatting system. It also incorporates capability for video selection and recording for assessment of the sensors monitored by the system. The system is software reprogrammable to numerous configurations that may be utilized for the collection of environmental, bi-level, analog and video data. The output of the system is digital tapes formatted for direct data reduction on a CDC 6400 computer, and video tapes containing timed tagged information that can be correlated with the digital data

  17. French experience in transient data collection and fatigue monitoring of PWR`s nuclear steam supply system; Experience francaise sur la comptabilisation des transitoires et la surveillance en fatigue des chaudieres REP

    Energy Technology Data Exchange (ETDEWEB)

    Sabaton, M.; Morilhat, P.; Savoldelli, D.; Genette, P.

    1995-10-01

    Electricite de France (EDF), the french national electricity company, is operating 54 standardized pressurizer water reactors. This about 500 reactor-years experience in nuclear stations operation and maintenance area has allowed EDF to develop its own strategy for monitoring of age-related degradations of NPP systems and components relevant for plant safety and reliability. After more than fifteen years of experience in regulatory transient data collection and seven years of successful fatigue monitoring prototypes experimentation, EDF decided to design a new system called SYSFAC (acronym for SYsteme de Surveillance en FAtigue de la Chaudiere) devoted to transient logging and thermal fatigue monitoring of the reactor coolant pressure boundary. The system is fully automatic and directly connected to the on-site data acquisition network without any complementary instrumentation. A functional transient detection module and a mechanical transient detection module are in charge of the general transient data collection. A fatigue monitoring module is aimed towards a precise surveillance of five specific zones particularly sensible to thermal fatigue. After the first step of preliminary studies, the industrial phase of the SYSFAC project is currently going on, with hardware and software tests and implementation. The first SYSFAC system will be delivered to the pilot power plant by the beginning of 1996. The extension to all EDF`s nuclear 900 MW is planned after one more year of feedback experience. (authors). 12 refs., 3 figs.

  18. ATIPS: Automatic Travel Itinerary Planning System for Domestic Areas.

    Science.gov (United States)

    Chang, Hsien-Tsung; Chang, Yi-Ming; Tsai, Meng-Tze

    2016-01-01

    Leisure travel has become a topic of great interest to Taiwanese residents in recent years. Most residents expect to be able to relax on a vacation during the holidays; however, the complicated procedure of travel itinerary planning is often discouraging and leads them to abandon the idea of traveling. In this paper, we design an automatic travel itinerary planning system for the domestic area (ATIPS) using an algorithm to automatically plan a domestic travel itinerary based on user intentions that allows users to minimize the process of trip planning. Simply by entering the travel time, the departure point, and the destination location, the system can automatically generate a travel itinerary. According to the results of the experiments, 70% of users were satisfied with the result of our system, and 82% of users were satisfied with the automatic user preference learning mechanism of ATIPS. Our algorithm also provides a framework for substituting modules or weights and offers a new method for travel planning.

  19. Approximate Sensory Data Collection: A Survey.

    Science.gov (United States)

    Cheng, Siyao; Cai, Zhipeng; Li, Jianzhong

    2017-03-10

    With the rapid development of the Internet of Things (IoTs), wireless sensor networks (WSNs) and related techniques, the amount of sensory data manifests an explosive growth. In some applications of IoTs and WSNs, the size of sensory data has already exceeded several petabytes annually, which brings too many troubles and challenges for the data collection, which is a primary operation in IoTs and WSNs. Since the exact data collection is not affordable for many WSN and IoT systems due to the limitations on bandwidth and energy, many approximate data collection algorithms have been proposed in the last decade. This survey reviews the state of the art of approximatedatacollectionalgorithms. Weclassifythemintothreecategories: themodel-basedones, the compressive sensing based ones, and the query-driven ones. For each category of algorithms, the advantages and disadvantages are elaborated, some challenges and unsolved problems are pointed out, and the research prospects are forecasted.

  20. Approximate Sensory Data Collection: A Survey

    Directory of Open Access Journals (Sweden)

    Siyao Cheng

    2017-03-01

    Full Text Available With the rapid development of the Internet of Things (IoTs, wireless sensor networks (WSNs and related techniques, the amount of sensory data manifests an explosive growth. In some applications of IoTs and WSNs, the size of sensory data has already exceeded several petabytes annually, which brings too many troubles and challenges for the data collection, which is a primary operation in IoTs and WSNs. Since the exact data collection is not affordable for many WSN and IoT systems due to the limitations on bandwidth and energy, many approximate data collection algorithms have been proposed in the last decade. This survey reviews the state of the art of approximatedatacollectionalgorithms. Weclassifythemintothreecategories: themodel-basedones, the compressive sensing based ones, and the query-driven ones. For each category of algorithms, the advantages and disadvantages are elaborated, some challenges and unsolved problems are pointed out, and the research prospects are forecasted.

  1. Automatic measurement system for light element isotope analysis

    International Nuclear Information System (INIS)

    Satake, Hiroshi; Ikegami, Kouichi.

    1990-01-01

    The automatic measurement system for the light element isotope analysis was developed by installing the specially designed inlet system which was controlled by a computer. The microcomputer system contains specific interface boards for the inlet system and the mass spectrometer, Micromass 602 E. All the components of the inlet and the computer system installed are easily available in Japan. Ten samples can be automatically measured as a maximum of. About 160 minutes are required for 10 measurements of δ 18 O values of CO 2 . Thus four samples can be measured per an hour using this system, while usually three samples for an hour using the manual operation. The automatized analysis system clearly has an advantage over the conventional method. This paper describes the details of this automated system, such as apparatuses used, the control procedure and the correction for reliable measurement. (author)

  2. MOLIERE: Automatic Biomedical Hypothesis Generation System.

    Science.gov (United States)

    Sybrandt, Justin; Shtutman, Michael; Safro, Ilya

    2017-08-01

    Hypothesis generation is becoming a crucial time-saving technique which allows biomedical researchers to quickly discover implicit connections between important concepts. Typically, these systems operate on domain-specific fractions of public medical data. MOLIERE, in contrast, utilizes information from over 24.5 million documents. At the heart of our approach lies a multi-modal and multi-relational network of biomedical objects extracted from several heterogeneous datasets from the National Center for Biotechnology Information (NCBI). These objects include but are not limited to scientific papers, keywords, genes, proteins, diseases, and diagnoses. We model hypotheses using Latent Dirichlet Allocation applied on abstracts found near shortest paths discovered within this network, and demonstrate the effectiveness of MOLIERE by performing hypothesis generation on historical data. Our network, implementation, and resulting data are all publicly available for the broad scientific community.

  3. Data collection architecture for big data - A framework for a research agenda

    NARCIS (Netherlands)

    Hofman, W.J.

    2015-01-01

    As big data is expected to contribute largely to economic growth, scalability of solutions becomes apparent for deployment by organisations. It requires automatic collection and processing of large, heterogeneous data sets of a variety of resources, dealing with various aspects like improving

  4. Automatic Emergence Detection in Complex Systems

    Directory of Open Access Journals (Sweden)

    Eugene Santos

    2017-01-01

    Full Text Available Complex systems consist of multiple interacting subsystems, whose nonlinear interactions can result in unanticipated (emergent system events. Extant systems analysis approaches fail to detect such emergent properties, since they analyze each subsystem separately and arrive at decisions typically through linear aggregations of individual analysis results. In this paper, we propose a quantitative definition of emergence for complex systems. We also propose a framework to detect emergent properties given observations of its subsystems. This framework, based on a probabilistic graphical model called Bayesian Knowledge Bases (BKBs, learns individual subsystem dynamics from data, probabilistically and structurally fuses said dynamics into a single complex system dynamics, and detects emergent properties. Fusion is the central element of our approach to account for situations when a common variable may have different probabilistic distributions in different subsystems. We evaluate our detection performance against a baseline approach (Bayesian Network ensemble on synthetic testbeds from UCI datasets. To do so, we also introduce a method to simulate and a metric to measure discrepancies that occur with shared/common variables. Experiments demonstrate that our framework outperforms the baseline. In addition, we demonstrate that this framework has uniform polynomial time complexity across all three learning, fusion, and reasoning procedures.

  5. Quasiinvariant Automatic Control Digital Systems of Inertia Objects

    OpenAIRE

    Lvov, Volodymyr; Andrieiev, Anatoliy

    2010-01-01

    The two-connected automatic control digital system (ACDS) and system of ACDS with combined control are examined. The two-connected and combined system of ACDS with work in the mode of tracking and stabilizing are analyzed. The discrete transfer function of two-connected and combined systems are obtained.

  6. An automatic evaluation system for NTA film neutron dosimeters

    CERN Document Server

    Müller, R

    1999-01-01

    At CERN, neutron personal monitoring for over 4000 collaborators is performed with Kodak NTA films, which have been shown to be the most suitable neutron dosimeter in the radiation environment around high-energy accelerators. To overcome the lengthy and strenuous manual scanning process with an optical microscope, an automatic analysis system has been developed. We report on the successful automatic scanning of NTA films irradiated with sup 2 sup 3 sup 8 Pu-Be source neutrons, which results in densely ionised recoil tracks, as well as on the extension of the method to higher energy neutrons causing sparse and fragmentary tracks. The application of the method in routine personal monitoring is discussed. $9 overcome the lengthy and strenuous manual scanning process with an optical microscope, an automatic analysis system has been developed. We report on the successful automatic scanning of NTA films irradiated with /sup 238/Pu-Be source $9 discussed. (10 refs).

  7. Automatic radiation measuring system connected with GPS

    International Nuclear Information System (INIS)

    Tanigaki, Minoru

    2014-01-01

    The most serious nuclear disaster in Japan has broken out at Fukushima Daiichi Nuclear Power Plant due to Great East Japan Earthquake. Prompt and exact mapping of the contamination is of great importance for radiation protection and for the environment restoration. We have developed radiation survey systems KURAMA and KURAMA-2 for rapid and exact measurement of radiation dose distribution. The system is composed of a mobile radiation monitor and the computer in office which is for the storage and visualization of the data. They are connected with internet and are operated for continuous radiation measurement while the monitor is moving. The mobile part consists of a survey meter, an interface to transform the output of the survey meter for the computer, a global positioning system, a computer to process the data for connecting to the network, and a mobile router. Thus they are effective for rapid mapping of the surface contamination. The operation and the performance of the equipment at the site are presented. (J.P.N.)

  8. An automatic system for acidity determination based on sequential injection titration and the monosegmented flow approach.

    Science.gov (United States)

    Kozak, Joanna; Wójtowicz, Marzena; Gawenda, Nadzieja; Kościelniak, Paweł

    2011-06-15

    An automatic sequential injection system, combining monosegmented flow analysis, sequential injection analysis and sequential injection titration is proposed for acidity determination. The system enables controllable sample dilution and generation of standards of required concentration in a monosegmented sequential injection manner, sequential injection titration of the prepared solutions, data collecting, and handling. It has been tested on spectrophotometric determination of acetic, citric and phosphoric acids with sodium hydroxide used as a titrant and phenolphthalein or thymolphthalein (in the case of phosphoric acid determination) as indicators. Accuracy better than |4.4|% (RE) and repeatability better than 2.9% (RSD) have been obtained. It has been applied to the determination of total acidity in vinegars and various soft drinks. The system provides low sample (less than 0.3 mL) consumption. On average, analysis of a sample takes several minutes. Copyright © 2011 Elsevier B.V. All rights reserved.

  9. Toward Global Comparability of Sexual Orientation Data in Official Statistics: A Conceptual Framework of Sexual Orientation for Health Data Collection in New Zealand’s Official Statistics System

    Directory of Open Access Journals (Sweden)

    Frank Pega

    2013-01-01

    Full Text Available Objective. Effectively addressing health disparities experienced by sexual minority populations requires high-quality official data on sexual orientation. We developed a conceptual framework of sexual orientation to improve the quality of sexual orientation data in New Zealand’s Official Statistics System. Methods. We reviewed conceptual and methodological literature, culminating in a draft framework. To improve the framework, we held focus groups and key-informant interviews with sexual minority stakeholders and producers and consumers of official statistics. An advisory board of experts provided additional guidance. Results. The framework proposes working definitions of the sexual orientation topic and measurement concepts, describes dimensions of the measurement concepts, discusses variables framing the measurement concepts, and outlines conceptual grey areas. Conclusion. The framework proposes standard definitions and concepts for the collection of official sexual orientation data in New Zealand. It presents a model for producers of official statistics in other countries, who wish to improve the quality of health data on their citizens.

  10. Water sample-collection and distribution system

    Science.gov (United States)

    Brooks, R. R.

    1978-01-01

    Collection and distribution system samples water from six designated stations, filtered if desired, and delivers it to various analytical sensors. System may be controlled by Water Monitoring Data Acquisition System or operated manually.

  11. Algorithm for automatic analysis of electro-oculographic data.

    Science.gov (United States)

    Pettersson, Kati; Jagadeesan, Sharman; Lukander, Kristian; Henelius, Andreas; Haeggström, Edward; Müller, Kiti

    2013-10-25

    Large amounts of electro-oculographic (EOG) data, recorded during electroencephalographic (EEG) measurements, go underutilized. We present an automatic, auto-calibrating algorithm that allows efficient analysis of such data sets. The auto-calibration is based on automatic threshold value estimation. Amplitude threshold values for saccades and blinks are determined based on features in the recorded signal. The performance of the developed algorithm was tested by analyzing 4854 saccades and 213 blinks recorded in two different conditions: a task where the eye movements were controlled (saccade task) and a task with free viewing (multitask). The results were compared with results from a video-oculography (VOG) device and manually scored blinks. The algorithm achieved 93% detection sensitivity for blinks with 4% false positive rate. The detection sensitivity for horizontal saccades was between 98% and 100%, and for oblique saccades between 95% and 100%. The classification sensitivity for horizontal and large oblique saccades (10 deg) was larger than 89%, and for vertical saccades larger than 82%. The duration and peak velocities of the detected horizontal saccades were similar to those in the literature. In the multitask measurement the detection sensitivity for saccades was 97% with a 6% false positive rate. The developed algorithm enables reliable analysis of EOG data recorded both during EEG and as a separate metrics.

  12. Improving labeling efficiency in automatic quality control of MRSI data.

    Science.gov (United States)

    Pedrosa de Barros, Nuno; McKinley, Richard; Wiest, Roland; Slotboom, Johannes

    2017-12-01

    To improve the efficiency of the labeling task in automatic quality control of MR spectroscopy imaging data. 28'432 short and long echo time (TE) spectra (1.5 tesla; point resolved spectroscopy (PRESS); repetition time (TR)= 1,500 ms) from 18 different brain tumor patients were labeled by two experts as either accept or reject, depending on their quality. For each spectrum, 47 signal features were extracted. The data was then used to run several simulations and test an active learning approach using uncertainty sampling. The performance of the classifiers was evaluated as a function of the number of patients in the training set, number of spectra in the training set, and a parameter α used to control the level of classification uncertainty required for a new spectrum to be selected for labeling. The results showed that the proposed strategy allows reductions of up to 72.97% for short TE and 62.09% for long TE in the amount of data that needs to be labeled, without significant impact in classification accuracy. Further reductions are possible with significant but minimal impact in performance. Active learning using uncertainty sampling is an effective way to increase the labeling efficiency for training automatic quality control classifiers. Magn Reson Med 78:2399-2405, 2017. © 2017 International Society for Magnetic Resonance in Medicine. © 2017 International Society for Magnetic Resonance in Medicine.

  13. System of acquisition and analysis of ultrasonic data

    International Nuclear Information System (INIS)

    Vaubert, Y.; Birac, A.M.; Saglio, R.

    1982-08-01

    An original system of acquisition and analysis of ultrasonic data collected during examinations named STADUS-PRODUS has been developed by C.E.A. in Saclay. First developed for the needs of in-service inspection of PWR vessels, it is now used for the different automatic ultrasonic controls with various tools

  14. Development of automatic cross section compilation system for MCNP

    International Nuclear Information System (INIS)

    Maekawa, Fujio; Sakurai, Kiyoshi

    1999-01-01

    A development of a code system to automatically convert cross-sections for MCNP is in progress. The NJOY code is, in general, used to convert the data compiled in the ENDF format (Evaluated Nuclear Data Files by BNL) into the cross-section libraries required by various reactor physics codes. While the cross-section library: FSXLIB-J3R2 was already converted from the JENDL-3.2 version of Japanese Evaluated Nuclear Data Library for a continuous energy Monte Carlo code MCNP, the library keeps only the cross-sections at room temperature (300 K). According to the users requirements which want to have cross-sections at higher temperature, say 600 K or 900 K, a code system named 'autonj' is under development to provide a set of cross-section library of arbitrary temperature for the MCNP code. This system can accept any of data formats adopted JENDL that may not be treated by NJOY code. The input preparation that is repeatedly required at every nuclide on NJOY execution is greatly reduced by permitting the conversion process of as many nuclides as the user wants in one execution. A few MCNP runs were achieved for verification purpose by using two libraries FSXLIB-J3R2 and the output of autonj'. The almost identical MCNP results within the statistical errors show the 'autonj' output library is correct. In FY 1998, the system will be completed, and in FY 1999, the user's manual will be published. (K. Tsuchihashi)

  15. Towards Automatic Music Transcription: Extraction of MIDI-Data out of Polyphonic Piano Music

    Directory of Open Access Journals (Sweden)

    Jens Wellhausen

    2005-06-01

    Full Text Available Driven by the increasing amount of music available electronically the need of automatic search and retrieval systems for music becomes more and more important. In this paper an algorithm for automatic transcription of polyphonic piano music into MIDI data is presented, which is a very interesting basis for database applications and music analysis. The first part of the algorithm performs a note accurate temporal audio segmentation. The resulting segments are examined to extract the notes played in the second part. An algorithm for chord separation based on Independent Subspace Analysis is presented. Finally, the results are used to build a MIDI file.

  16. Simultaneous optical and meteor head echo measurements using the Middle Atmosphere Alomar Radar System (MAARSY): Data collection and preliminary analysis

    Science.gov (United States)

    Brown, P.; Stober, G.; Schult, C.; Krzeminski, Z.; Cooke, W.; Chau, J. L.

    2017-07-01

    The initial results of a two year simultaneous optical-radar meteor campaign are described. Analysis of 105 double-station optical meteors having plane of sky intersection angles greater than 5° and trail lengths in excess of 2 km also detected by the Middle Atmosphere Alomar Radar System (MAARSY) as head echoes was performed. These events show a median deviation in radiants between radar and optical determinations of 1.5°, with 1/3 of events having radiant agreement to less than one degree. MAARSY tends to record average speeds roughly 0.5 km/s and 1.3 km higher than optical records, in part due to the higher sensitivity of MAARSY as compared to the optical instruments. More than 98% of all head echoes are not detected with the optical system. Using this non-detection ratio and the known limiting sensitivity of the cameras, we estimate that the limiting meteoroid detection mass of MAARSY is in the 10-9-10-10 kg (astronomical limiting meteor magnitudes of +11 to +12) appropriate to speeds from 30 to 60 km/s. There is a clear trend of higher peak RCS for brighter meteors between 35 and -30 dBsm. For meteors with similar magnitudes, the MAARSY head echo radar cross-section is larger at higher speeds. Brighter meteors at fixed heights and similar speeds have consistently, on average, larger RCS values, in accordance with established scattering theory. However, our data show RCS ∝ v/2, much weaker than the normally assumed RCS ∝ v3, a consequence of our requiring head echoes to also be detectable optically. Most events show a smooth variation of RCS with height broadly following the light production behavior. A significant minority of meteors show large variations in RCS relative to the optical light curve over common height intervals, reflecting fragmentation or possibly differential ablation. No optically detected meteor occurring in the main radar beam and at times when the radar was collecting head echo data went unrecorded by MAARSY. Thus there does not

  17. A System for Automatically Generating Scheduling Heuristics

    Science.gov (United States)

    Morris, Robert

    1996-01-01

    The goal of this research is to improve the performance of automated schedulers by designing and implementing an algorithm by automatically generating heuristics by selecting a schedule. The particular application selected by applying this method solves the problem of scheduling telescope observations, and is called the Associate Principal Astronomer. The input to the APA scheduler is a set of observation requests submitted by one or more astronomers. Each observation request specifies an observation program as well as scheduling constraints and preferences associated with the program. The scheduler employs greedy heuristic search to synthesize a schedule that satisfies all hard constraints of the domain and achieves a good score with respect to soft constraints expressed as an objective function established by an astronomer-user.

  18. AUTOMATIC 3D BUILDING MODEL GENERATIONS WITH AIRBORNE LiDAR DATA

    Directory of Open Access Journals (Sweden)

    N. Yastikli

    2017-11-01

    Full Text Available LiDAR systems become more and more popular because of the potential use for obtaining the point clouds of vegetation and man-made objects on the earth surface in an accurate and quick way. Nowadays, these airborne systems have been frequently used in wide range of applications such as DEM/DSM generation, topographic mapping, object extraction, vegetation mapping, 3 dimensional (3D modelling and simulation, change detection, engineering works, revision of maps, coastal management and bathymetry. The 3D building model generation is the one of the most prominent applications of LiDAR system, which has the major importance for urban planning, illegal construction monitoring, 3D city modelling, environmental simulation, tourism, security, telecommunication and mobile navigation etc. The manual or semi-automatic 3D building model generation is costly and very time-consuming process for these applications. Thus, an approach for automatic 3D building model generation is needed in a simple and quick way for many studies which includes building modelling. In this study, automatic 3D building models generation is aimed with airborne LiDAR data. An approach is proposed for automatic 3D building models generation including the automatic point based classification of raw LiDAR point cloud. The proposed point based classification includes the hierarchical rules, for the automatic production of 3D building models. The detailed analyses for the parameters which used in hierarchical rules have been performed to improve classification results using different test areas identified in the study area. The proposed approach have been tested in the study area which has partly open areas, forest areas and many types of the buildings, in Zekeriyakoy, Istanbul using the TerraScan module of TerraSolid. The 3D building model was generated automatically using the results of the automatic point based classification. The obtained results of this research on study area verified

  19. Automatic 3d Building Model Generations with Airborne LiDAR Data

    Science.gov (United States)

    Yastikli, N.; Cetin, Z.

    2017-11-01

    LiDAR systems become more and more popular because of the potential use for obtaining the point clouds of vegetation and man-made objects on the earth surface in an accurate and quick way. Nowadays, these airborne systems have been frequently used in wide range of applications such as DEM/DSM generation, topographic mapping, object extraction, vegetation mapping, 3 dimensional (3D) modelling and simulation, change detection, engineering works, revision of maps, coastal management and bathymetry. The 3D building model generation is the one of the most prominent applications of LiDAR system, which has the major importance for urban planning, illegal construction monitoring, 3D city modelling, environmental simulation, tourism, security, telecommunication and mobile navigation etc. The manual or semi-automatic 3D building model generation is costly and very time-consuming process for these applications. Thus, an approach for automatic 3D building model generation is needed in a simple and quick way for many studies which includes building modelling. In this study, automatic 3D building models generation is aimed with airborne LiDAR data. An approach is proposed for automatic 3D building models generation including the automatic point based classification of raw LiDAR point cloud. The proposed point based classification includes the hierarchical rules, for the automatic production of 3D building models. The detailed analyses for the parameters which used in hierarchical rules have been performed to improve classification results using different test areas identified in the study area. The proposed approach have been tested in the study area which has partly open areas, forest areas and many types of the buildings, in Zekeriyakoy, Istanbul using the TerraScan module of TerraSolid. The 3D building model was generated automatically using the results of the automatic point based classification. The obtained results of this research on study area verified that automatic 3D

  20. Human visual system automatically encodes sequential regularities of discrete events.

    Science.gov (United States)

    Kimura, Motohiro; Schröger, Erich; Czigler, István; Ohira, Hideki

    2010-06-01

    For our adaptive behavior in a dynamically changing environment, an essential task of the brain is to automatically encode sequential regularities inherent in the environment into a memory representation. Recent studies in neuroscience have suggested that sequential regularities embedded in discrete sensory events are automatically encoded into a memory representation at the level of the sensory system. This notion is largely supported by evidence from investigations using auditory mismatch negativity (auditory MMN), an event-related brain potential (ERP) correlate of an automatic memory-mismatch process in the auditory sensory system. However, it is still largely unclear whether or not this notion can be generalized to other sensory modalities. The purpose of the present study was to investigate the contribution of the visual sensory system to the automatic encoding of sequential regularities using visual mismatch negativity (visual MMN), an ERP correlate of an automatic memory-mismatch process in the visual sensory system. To this end, we conducted a sequential analysis of visual MMN in an oddball sequence consisting of infrequent deviant and frequent standard stimuli, and tested whether the underlying memory representation of visual MMN generation contains only a sensory memory trace of standard stimuli (trace-mismatch hypothesis) or whether it also contains sequential regularities extracted from the repetitive standard sequence (regularity-violation hypothesis). The results showed that visual MMN was elicited by first deviant (deviant stimuli following at least one standard stimulus), second deviant (deviant stimuli immediately following first deviant), and first standard (standard stimuli immediately following first deviant), but not by second standard (standard stimuli immediately following first standard). These results are consistent with the regularity-violation hypothesis, suggesting that the visual sensory system automatically encodes sequential

  1. An automatic tension measurement system of MWPC wires

    International Nuclear Information System (INIS)

    D'Antone, I.; Lolli, M.; Torromeo, G.

    1992-01-01

    An electronic system is presented for automatic mechanical tension measurement to test wire chambers. The developed system works in the tension range from 50 g to 300 g; this large working range is obtained by using a microcontroller that performs a digital control on the bridge of an oscillator containing the wire of which the tension has to be measured. The microcontroller automatically brings the system towards the oscillation condition and subsequently, measuring the frequency, it evaluates, displays and sends to a host computer the value of the mechanical tension of the wires. The system is precise and allows fast measurements. A description of the hardware and software design is given. (orig.)

  2. Longline Observer Data System

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — LODS, the Hawaii Longline Observer Data System, is a complete suite of tools designed to collect, process, and manage quality fisheries data and information. Guided...

  3. Signal system data mining

    Science.gov (United States)

    2000-09-01

    Intelligent transportation systems (ITS) include large numbers of traffic sensors that collect enormous quantities of data. The data provided by ITS is necessary for advanced forms of control, however basic forms of control, primarily time-of-day (TO...

  4. Automatic Road Centerline Extraction from Imagery Using Road GPS Data

    Directory of Open Access Journals (Sweden)

    Chuqing Cao

    2014-09-01

    Full Text Available Road centerline extraction from imagery constitutes a key element in numerous geospatial applications, which has been addressed through a variety of approaches. However, most of the existing methods are not capable of dealing with challenges such as different road shapes, complex scenes, and variable resolutions. This paper presents a novel method for road centerline extraction from imagery in a fully automatic approach that addresses the aforementioned challenges by exploiting road GPS data. The proposed method combines road color feature with road GPS data to detect road centerline seed points. After global alignment of road GPS data, a novel road centerline extraction algorithm is developed to extract each individual road centerline in local regions. Through road connection, road centerline network is generated as the final output. Extensive experiments demonstrate that our proposed method can rapidly and accurately extract road centerline from remotely sensed imagery.

  5. An inexpensive compact automatic camera system for wildlife research

    Science.gov (United States)

    William R. Danielson; Richard M. DeGraaf; Todd K. Fuller

    1996-01-01

    This paper describes the design, conversion, and deployment of a reliable, compact, automatic multiple-exposure photographic system that was used to photograph nest predation events. This system may be the most versatile yet described in the literature because of its simplicity, portability, and dependability. The system was very reliable because it was designed around...

  6. Automatic Management of Parallel and Distributed System Resources

    Science.gov (United States)

    Yan, Jerry; Ngai, Tin Fook; Lundstrom, Stephen F.

    1990-01-01

    Viewgraphs on automatic management of parallel and distributed system resources are presented. Topics covered include: parallel applications; intelligent management of multiprocessing systems; performance evaluation of parallel architecture; dynamic concurrent programs; compiler-directed system approach; lattice gaseous cellular automata; and sparse matrix Cholesky factorization.

  7. 29 CFR 1910.159 - Automatic sprinkler systems.

    Science.gov (United States)

    2010-07-01

    ... supply is out of service, except for systems of 20 or fewer sprinklers. (5) Hose connections for fire fighting use. The employer may attach hose connections for fire fighting use to wet pipe sprinkler systems... 29 Labor 5 2010-07-01 2010-07-01 false Automatic sprinkler systems. 1910.159 Section 1910.159...

  8. AUTOMATIC CONTROL SYSTEM ОF REGIONAL BOILER HOUSE

    Directory of Open Access Journals (Sweden)

    V. A. Sednin

    2005-01-01

    Full Text Available Implementation of an automatic control system is one of directions that promotes to increase an operational efficiency of a heat supply system. A heating boiler house in Surgut (Russia is taken as an example to demonstrate an actual realization of such system.

  9. THE AUTOMATIC LIGHTENING LOCATION SYSTEM AND ITS ...

    African Journals Online (AJOL)

    ES Obe

    systems. The implications of the lightening location system for the Nigerian electric power system are also highlighted. ... system (the LLP type) is currently operating in may countries .... (iii) Real time lightning maps will aid service restoration.

  10. Formal Specification Based Automatic Test Generation for Embedded Network Systems

    Directory of Open Access Journals (Sweden)

    Eun Hye Choi

    2014-01-01

    Full Text Available Embedded systems have become increasingly connected and communicate with each other, forming large-scaled and complicated network systems. To make their design and testing more reliable and robust, this paper proposes a formal specification language called SENS and a SENS-based automatic test generation tool called TGSENS. Our approach is summarized as follows: (1 A user describes requirements of target embedded network systems by logical property-based constraints using SENS. (2 Given SENS specifications, test cases are automatically generated using a SAT-based solver. Filtering mechanisms to select efficient test cases are also available in our tool. (3 In addition, given a testing goal by the user, test sequences are automatically extracted from exhaustive test cases. We’ve implemented our approach and conducted several experiments on practical case studies. Through the experiments, we confirmed the efficiency of our approach in design and test generation of real embedded air-conditioning network systems.

  11. Temporally rendered automatic cloud extraction (TRACE) system

    Science.gov (United States)

    Bodrero, Dennis M.; Yale, James G.; Davis, Roger E.; Rollins, John M.

    1999-10-01

    Smoke/obscurant testing requires that 2D cloud extent be extracted from visible and thermal imagery. These data are used alone or in combination with 2D data from other aspects to make 3D calculations of cloud properties, including dimensions, volume, centroid, travel, and uniformity. Determining cloud extent from imagery has historically been a time-consuming manual process. To reduce time and cost associated with smoke/obscurant data processing, automated methods to extract cloud extent from imagery were investigated. The TRACE system described in this paper was developed and implemented at U.S. Army Dugway Proving Ground, UT by the Science and Technology Corporation--Acuity Imaging Incorporated team with Small Business Innovation Research funding. TRACE uses dynamic background subtraction and 3D fast Fourier transform as primary methods to discriminate the smoke/obscurant cloud from the background. TRACE has been designed to run on a PC-based platform using Windows. The PC-Windows environment was chosen for portability, to give TRACE the maximum flexibility in terms of its interaction with peripheral hardware devices such as video capture boards, removable media drives, network cards, and digital video interfaces. Video for Windows provides all of the necessary tools for the development of the video capture utility in TRACE and allows for interchangeability of video capture boards without any software changes. TRACE is designed to take advantage of future upgrades in all aspects of its component hardware. A comparison of cloud extent determined by TRACE with manual method is included in this paper.

  12. Automatic multi-camera calibration for deployable positioning systems

    Science.gov (United States)

    Axelsson, Maria; Karlsson, Mikael; Rudner, Staffan

    2012-06-01

    Surveillance with automated positioning and tracking of subjects and vehicles in 3D is desired in many defence and security applications. Camera systems with stereo or multiple cameras are often used for 3D positioning. In such systems, accurate camera calibration is needed to obtain a reliable 3D position estimate. There is also a need for automated camera calibration to facilitate fast deployment of semi-mobile multi-camera 3D positioning systems. In this paper we investigate a method for automatic calibration of the extrinsic camera parameters (relative camera pose and orientation) of a multi-camera positioning system. It is based on estimation of the essential matrix between each camera pair using the 5-point method for intrinsically calibrated cameras. The method is compared to a manual calibration method using real HD video data from a field trial with a multicamera positioning system. The method is also evaluated on simulated data from a stereo camera model. The results show that the reprojection error of the automated camera calibration method is close to or smaller than the error for the manual calibration method and that the automated calibration method can replace the manual calibration.

  13. Automatic generation of configuration files for a distributed control system

    CERN Document Server

    Cupérus, J

    1995-01-01

    The CERN PS accelerator complex is composed of 9 interlinked accelerators for production and acceleration of various kinds of particles. The hardware is controlled through CAMAC, VME, G64, and GPIB modules, which in turn are controlled by more than 100 microprocessors in VME crates. To produce startup files for all these microprocessors, with the correct drivers, programs and parameters in each of them, is quite a challenge. The problem is solved by generating the startup files automatically from the description of the control system in a relational database. The generation process detects inconsistencies and incomplete information. Included in the startup files are data which are formally comments, but can be interpreted for run-time checking of interface modules and program activity.

  14. Channel Access Algorithm Design for Automatic Identification System

    Institute of Scientific and Technical Information of China (English)

    Oh Sang-heon; Kim Seung-pum; Hwang Dong-hwan; Park Chan-sik; Lee Sang-jeong

    2003-01-01

    The Automatic Identification System (AIS) is a maritime equipment to allow an efficient exchange of the navigational data between ships and between ships and shore stations. It utilizes a channel access algorithm which can quickly resolve conflicts without any intervention from control stations. In this paper, a design of channel access algorithm for the AIS is presented. The input/output relationship of each access algorithm module is defined by drawing the state transition diagram, dataflow diagram and flowchart based on the technical standard, ITU-R M.1371. In order to verify the designed channel access algorithm, the simulator was developed using the C/C++ programming language. The results show that the proposed channel access algorithm can properly allocate transmission slots and meet the operational performance requirements specified by the technical standard.

  15. The Siegen automatic measuring system for track detectors: new developments

    International Nuclear Information System (INIS)

    Rusch, G.; Winkel, E.; Noll, A.; Heinrich, W.

    1991-01-01

    Starting twelve years ago we have developed completely automatic scanning and measuring systems for nuclear track detectors. The hardware and software of these systems have continuously been improved. They were used in different heavy ion and cosmic ray experiments. In this paper we describe methods for high resolution REL measurements in plastic nuclear track detectors and methods to scan and measure nuclear disintegration stars in AgCl detectors using an automatic measuring technique. The system uses a stepping motor driven microscope stage, a video camera and an image analysis computer based on a MC68020 microprocessor. (author)

  16. Automatic digital surface model (DSM) generation from aerial imagery data

    Science.gov (United States)

    Zhou, Nan; Cao, Shixiang; He, Hongyan; Xing, Kun; Yue, Chunyu

    2018-04-01

    Aerial sensors are widely used to acquire imagery for photogrammetric and remote sensing application. In general, the images have large overlapped region, which provide a lot of redundant geometry and radiation information for matching. This paper presents a POS supported dense matching procedure for automatic DSM generation from aerial imagery data. The method uses a coarse-to-fine hierarchical strategy with an effective combination of several image matching algorithms: image radiation pre-processing, image pyramid generation, feature point extraction and grid point generation, multi-image geometrically constraint cross-correlation (MIG3C), global relaxation optimization, multi-image geometrically constrained least squares matching (MIGCLSM), TIN generation and point cloud filtering. The image radiation pre-processing is used in order to reduce the effects of the inherent radiometric problems and optimize the images. The presented approach essentially consists of 3 components: feature point extraction and matching procedure, grid point matching procedure and relational matching procedure. The MIGCLSM method is used to achieve potentially sub-pixel accuracy matches and identify some inaccurate and possibly false matches. The feasibility of the method has been tested on different aerial scale images with different landcover types. The accuracy evaluation is based on the comparison between the automatic extracted DSMs derived from the precise exterior orientation parameters (EOPs) and the POS.

  17. An automatic counting and recording system (1963); Ensemble de comptage a enregistrement automatique (1963)

    Energy Technology Data Exchange (ETDEWEB)

    Pierre, B. [Commissariat a l' Energie Atomique, Saclay (France). Centre d' Etudes Nucleaires

    1961-09-15

    An automatic control, counting and programing system for the collection of single crystal diffractometry data was designed by the author for a neutron diffractometer in 1958 at C.E.N - Grenoble. A part of the whole instrument, 'The Automatic Counting and Recording System', is described in this paper. Its applications are numerous and extensive, e.g.: the system has been designed for neutron diffractometer, but it can easily be adapted either for use with X-rays or measurement of mean life in {beta} decay analysis. (author) [French] Un ensemble automatique de telecommande, comptage et programmation pour la diffractometrie a cristal unique a ete etudie et realise par l'auteur pour la diffraction des neutrons en 1958 au C.E.N - Grenoble. Le present rapport decrit a ''l'Ensemble de Comptage a Enregistrement Automatique'' qui est une partie de l'appareillage complet. Ses applications sont nombreuses et peuvent s'etendre a de nouveaux domaines. En effet cet ensemble qui a ete etudie pour fonctionner avec un diffractometre a neutron, peut facilement s'adapter a la technique de diffraction des rayons X ou par exemple a celle de decroisasnce d'activite {beta}. (auteur)

  18. An automatic counting and recording system (1963); Ensemble de comptage a enregistrement automatique (1963)

    Energy Technology Data Exchange (ETDEWEB)

    Pierre, B [Commissariat a l' Energie Atomique, Saclay (France). Centre d' Etudes Nucleaires

    1961-09-15

    An automatic control, counting and programing system for the collection of single crystal diffractometry data was designed by the author for a neutron diffractometer in 1958 at C.E.N - Grenoble. A part of the whole instrument, 'The Automatic Counting and Recording System', is described in this paper. Its applications are numerous and extensive, e.g.: the system has been designed for neutron diffractometer, but it can easily be adapted either for use with X-rays or measurement of mean life in {beta} decay analysis. (author) [French] Un ensemble automatique de telecommande, comptage et programmation pour la diffractometrie a cristal unique a ete etudie et realise par l'auteur pour la diffraction des neutrons en 1958 au C.E.N - Grenoble. Le present rapport decrit a ''l'Ensemble de Comptage a Enregistrement Automatique'' qui est une partie de l'appareillage complet. Ses applications sont nombreuses et peuvent s'etendre a de nouveaux domaines. En effet cet ensemble qui a ete etudie pour fonctionner avec un diffractometre a neutron, peut facilement s'adapter a la technique de diffraction des rayons X ou par exemple a celle de decroisasnce d'activite {beta}. (auteur)

  19. Development of an Automatic Identification System Autonomous Positioning System

    Directory of Open Access Journals (Sweden)

    Qing Hu

    2015-11-01

    Full Text Available In order to overcome the vulnerability of the global navigation satellite system (GNSS and provide robust position, navigation and time (PNT information in marine navigation, the autonomous positioning system based on ranging-mode Automatic Identification System (AIS is presented in the paper. The principle of the AIS autonomous positioning system (AAPS is investigated, including the position algorithm, the signal measurement technique, the geometric dilution of precision, the time synchronization technique and the additional secondary factor correction technique. In order to validate the proposed AAPS, a verification system has been established in the Xinghai sea region of Dalian (China. Static and dynamic positioning experiments are performed. The original function of the AIS in the AAPS is not influenced. The experimental results show that the positioning precision of the AAPS is better than 10 m in the area with good geometric dilution of precision (GDOP by the additional secondary factor correction technology. This is the most economical solution for a land-based positioning system to complement the GNSS for the navigation safety of vessels sailing along coasts.

  20. Neuro-fuzzy system modeling based on automatic fuzzy clustering

    Institute of Scientific and Technical Information of China (English)

    Yuangang TANG; Fuchun SUN; Zengqi SUN

    2005-01-01

    A neuro-fuzzy system model based on automatic fuzzy clustering is proposed.A hybrid model identification algorithm is also developed to decide the model structure and model parameters.The algorithm mainly includes three parts:1) Automatic fuzzy C-means (AFCM),which is applied to generate fuzzy rules automatically,and then fix on the size of the neuro-fuzzy network,by which the complexity of system design is reducesd greatly at the price of the fitting capability;2) Recursive least square estimation (RLSE).It is used to update the parameters of Takagi-Sugeno model,which is employed to describe the behavior of the system;3) Gradient descent algorithm is also proposed for the fuzzy values according to the back propagation algorithm of neural network.Finally,modeling the dynamical equation of the two-link manipulator with the proposed approach is illustrated to validate the feasibility of the method.

  1. Automatic behaviour analysis system for honeybees using computer vision

    DEFF Research Database (Denmark)

    Tu, Gang Jun; Hansen, Mikkel Kragh; Kryger, Per

    2016-01-01

    We present a fully automatic online video system, which is able to detect the behaviour of honeybees at the beehive entrance. Our monitoring system focuses on observing the honeybees as naturally as possible (i.e. without disturbing the honeybees). It is based on the Raspberry Pi that is a low...

  2. Automatic design of optical systems by digital computer

    Science.gov (United States)

    Casad, T. A.; Schmidt, L. F.

    1967-01-01

    Computer program uses geometrical optical techniques and a least squares optimization method employing computing equipment for the automatic design of optical systems. It evaluates changes in various optical parameters, provides comprehensive ray-tracing, and generally determines the acceptability of the optical system characteristics.

  3. Automatic patient respiration failure detection system with wireless transmission

    Science.gov (United States)

    Dimeff, J.; Pope, J. M.

    1968-01-01

    Automatic respiration failure detection system detects respiration failure in patients with a surgically implanted tracheostomy tube, and actuates an audible and/or visual alarm. The system incorporates a miniature radio transmitter so that the patient is unencumbered by wires yet can be monitored from a remote location.

  4. Full-automatic Special Drill Hydraulic System and PLC Control

    Directory of Open Access Journals (Sweden)

    Tian Xue Jun

    2016-01-01

    Full Text Available A hydraulic-driven and PLC full-automatic special drill is introduced, working principle of the hydraulic system and PLC control system are analyzed and designed, this equipment has the advantages of high efficiency, superior quality and low cost etc.

  5. Robust Fallback Scheme for the Danish Automatic Voltage Control System

    DEFF Research Database (Denmark)

    Qin, Nan; Dmitrova, Evgenia; Lund, Torsten

    2015-01-01

    This paper proposes a fallback scheme for the Danish automatic voltage control system. It will be activated in case of the local station loses telecommunication to the control center and/or the local station voltage violates the acceptable operational limits. It cuts in/out switchable and tap...... power system....

  6. Evaluation of the SYSTRAN Automatic Translation System. Report No. 5.

    Science.gov (United States)

    Chaumier, Jacques; And Others

    The Commission of the European Communities has acquired an automatic translation system (SYSTRAN), which has been put into operation on an experimental basis. The system covers translation of English into French and comprises a dictionary for food science and technology containing 25,000 words or inflections and 4,500 expressions. This report…

  7. Text Structuration Leading to an Automatic Summary System: RAFI.

    Science.gov (United States)

    Lehman, Abderrafih

    1999-01-01

    Describes the design and construction of Resume Automatique a Fragments Indicateurs (RAFI), a system of automatic text summary which sums up scientific and technical texts. The RAFI system transforms a long source text into several versions of more condensed texts, using discourse analysis, to make searching easier; it could be adapted to the…

  8. Development of a computer-assisted personal interview software system for collection of tribal fish consumption data.

    Science.gov (United States)

    Kissinger, Lon; Lorenzana, Roseanne; Mittl, Beth; Lasrado, Merwyn; Iwenofu, Samuel; Olivo, Vanessa; Helba, Cynthia; Capoeman, Pauline; Williams, Ann H

    2010-12-01

    The authors developed a computer-assisted personal interviewing (CAPI) seafood consumption survey tool from existing Pacific NW Native American seafood consumption survey methodology. The software runs on readily available hardware and software, and is easily configured for different cultures and seafood resources. The CAPI is used with a booklet of harvest location maps and species and portion size images. The use of a CAPI facilitates tribal administration of seafood consumption surveys, allowing cost-effective collection of scientifically defensible data and tribal management of data and data interpretation. Use of tribal interviewers reduces potential bias and discomfort that may be associated with nontribal interviewers. The CAPI contains a 24-hour recall and food frequency questionnaire, and assesses seasonal seafood consumption and temporal changes in consumption. EPA's methodology for developing ambient water quality criteria for tribes assigns a high priority to local data. The CAPI will satisfy this guidance objective. Survey results will support development of tribal water quality standards on their lands and assessment of seafood consumption-related contaminant risks and nutritional benefits. CAPI advantages over paper surveys include complex question branching without raising respondent burden, more complete interviews due to answer error and range checking, data transcription error elimination, printing and mailing cost elimination, and improved data storage. The survey instrument was pilot tested among the Quinault Nation in 2006. © 2010 Society for Risk Analysis.

  9. Development of advanced automatic control system for nuclear ship. 2. Perfect automatic operation after reactor scram events

    International Nuclear Information System (INIS)

    Yabuuchi, Noriaki; Nakazawa, Toshio; Takahashi, Hiroki; Shimazaki, Junya; Hoshi, Tsutao

    1997-11-01

    An automatic operation system has been developed for the purpose of realizing a perfect automatic plant operation after reactor scram events. The goal of the automatic operation after a reactor scram event is to bring the reactor hot stand-by condition automatically. The basic functions of this system are as follows; to monitor actions of the equipments of safety actions after a reactor scram, to control necessary control equipments to bring a reactor to a hot stand-by condition automatically, and to energize a decay heat removal system. The performance evaluation on this system was carried out by comparing the results using to Nuclear Ship Engineering Simulation System (NESSY) and the those measured in the scram test of the nuclear ship 'Mutsu'. As the result, it was showed that this system had the sufficient performance to bring a reactor to a hot syand-by condition quickly and safety. (author)

  10. Development of advanced automatic control system for nuclear ship. 2. Perfect automatic operation after reactor scram events

    Energy Technology Data Exchange (ETDEWEB)

    Yabuuchi, Noriaki; Nakazawa, Toshio; Takahashi, Hiroki; Shimazaki, Junya; Hoshi, Tsutao [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1997-11-01

    An automatic operation system has been developed for the purpose of realizing a perfect automatic plant operation after reactor scram events. The goal of the automatic operation after a reactor scram event is to bring the reactor hot stand-by condition automatically. The basic functions of this system are as follows; to monitor actions of the equipments of safety actions after a reactor scram, to control necessary control equipments to bring a reactor to a hot stand-by condition automatically, and to energize a decay heat removal system. The performance evaluation on this system was carried out by comparing the results using to Nuclear Ship Engineering Simulation System (NESSY) and the those measured in the scram test of the nuclear ship `Mutsu`. As the result, it was showed that this system had the sufficient performance to bring a reactor to a hot syand-by condition quickly and safety. (author)

  11. Automated system for data acquisition and monitoring

    Directory of Open Access Journals (Sweden)

    Borza Sorin

    2017-01-01

    Full Text Available The Environmental management has become, with the development of human society a very important issue. There have been multiple systems that automatically monitors the environment. In this paper we propose a system that integrates GIS software and data acquisition software. In addition the proposed system implements new AHP multicriteria method that can get an answer online on each pollutant influence on limited geographical area in which the monitors. Factors pollutants of limited geographical areas are taken automatically by specific sensors through acquisition board. Labview software, with virtual instrument created by transferring them into a database Access. Access database they are taken up by software Geomedia Professional and processed using multi-criteria method AHP, so that at any moment, their influence on the environment and classify these influences, can be plotted on the screen monitoring system. The system allows, the automatic collection of data, the memorization and the generation of GIS elements. The research presented in this paper were aimed at implementing multi-criteria methods in GIS software.

  12. Accelerometer-controlled automatic braking system

    Science.gov (United States)

    Dreher, R. C.; Sleeper, R. K.; Nayadley, J. R., Sr.

    1973-01-01

    Braking system, which employs angular accelerometer to control wheel braking and results in low level of tire slip, has been developed and tested. Tests indicate that system is feasible for operations on surfaces of different slipperinesses. System restricts tire slip and is capable of adapting to rapidly-changing surface conditions.

  13. Single Point Vulnerability Analysis of Automatic Seismic Trip System

    Energy Technology Data Exchange (ETDEWEB)

    Oh, Seo Bin; Chung, Soon Il; Lee, Yong Suk [FNC Technology Co., Yongin (Korea, Republic of); Choi, Byung Pil [KHNP CRI, Daejeon (Korea, Republic of)

    2016-10-15

    Single Point Vulnerability (SPV) analysis is a process used to identify individual equipment whose failure alone will result in a reactor trip, turbine generator failure, or power reduction of more than 50%. Automatic Seismic Trip System (ASTS) is a newly installed system to ensure the safety of plant when earthquake occurs. Since this system directly shuts down the reactor, the failure or malfunction of its system component can cause a reactor trip more frequently than other systems. Therefore, an SPV analysis of ASTS is necessary to maintain its essential performance. To analyze SPV for ASTS, failure mode and effect analysis (FMEA) and fault tree analysis (FTA) was performed. In this study, FMEA and FTA methods were performed to select SPV equipment of ASTS. D/O, D/I, A/I card, seismic sensor, and trip relay had an effect on the reactor trip but their single failure will not cause reactor trip. In conclusion, ASTS is excluded as SPV. These results can be utilized as the basis data for ways to enhance facility reliability such as design modification and improvement of preventive maintenance procedure.

  14. Single Point Vulnerability Analysis of Automatic Seismic Trip System

    International Nuclear Information System (INIS)

    Oh, Seo Bin; Chung, Soon Il; Lee, Yong Suk; Choi, Byung Pil

    2016-01-01

    Single Point Vulnerability (SPV) analysis is a process used to identify individual equipment whose failure alone will result in a reactor trip, turbine generator failure, or power reduction of more than 50%. Automatic Seismic Trip System (ASTS) is a newly installed system to ensure the safety of plant when earthquake occurs. Since this system directly shuts down the reactor, the failure or malfunction of its system component can cause a reactor trip more frequently than other systems. Therefore, an SPV analysis of ASTS is necessary to maintain its essential performance. To analyze SPV for ASTS, failure mode and effect analysis (FMEA) and fault tree analysis (FTA) was performed. In this study, FMEA and FTA methods were performed to select SPV equipment of ASTS. D/O, D/I, A/I card, seismic sensor, and trip relay had an effect on the reactor trip but their single failure will not cause reactor trip. In conclusion, ASTS is excluded as SPV. These results can be utilized as the basis data for ways to enhance facility reliability such as design modification and improvement of preventive maintenance procedure

  15. Vital Recorder-a free research tool for automatic recording of high-resolution time-synchronised physiological data from multiple anaesthesia devices.

    Science.gov (United States)

    Lee, Hyung-Chul; Jung, Chul-Woo

    2018-01-24

    The current anaesthesia information management system (AIMS) has limited capability for the acquisition of high-quality vital signs data. We have developed a Vital Recorder program to overcome the disadvantages of AIMS and to support research. Physiological data of surgical patients were collected from 10 operating rooms using the Vital Recorder. The basic equipment used were a patient monitor, the anaesthesia machine, and the bispectral index (BIS) monitor. Infusion pumps, cardiac output monitors, regional oximeter, and rapid infusion device were added as required. The automatic recording option was used exclusively and the status of recording was frequently checked through web monitoring. Automatic recording was successful in 98.5% (4,272/4,335) cases during eight months of operation. The total recorded time was 13,489 h (3.2 ± 1.9 h/case). The Vital Recorder's automatic recording and remote monitoring capabilities enabled us to record physiological big data with minimal effort. The Vital Recorder also provided time-synchronised data captured from a variety of devices to facilitate an integrated analysis of vital signs data. The free distribution of the Vital Recorder is expected to improve data access for researchers attempting physiological data studies and to eliminate inequalities in research opportunities due to differences in data collection capabilities.

  16. Automatic User Interface Generation for Visualizing Big Geoscience Data

    Science.gov (United States)

    Yu, H.; Wu, J.; Zhou, Y.; Tang, Z.; Kuo, K. S.

    2016-12-01

    Along with advanced computing and observation technologies, geoscience and its related fields have been generating a large amount of data at an unprecedented growth rate. Visualization becomes an increasingly attractive and feasible means for researchers to effectively and efficiently access and explore data to gain new understandings and discoveries. However, visualization has been challenging due to a lack of effective data models and visual representations to tackle the heterogeneity of geoscience data. We propose a new geoscience data visualization framework by leveraging the interface automata theory to automatically generate user interface (UI). Our study has the following three main contributions. First, geoscience data has its unique hierarchy data structure and complex formats, and therefore it is relatively easy for users to get lost or confused during their exploration of the data. By applying interface automata model to the UI design, users can be clearly guided to find the exact visualization and analysis that they want. In addition, from a development perspective, interface automaton is also easier to understand than conditional statements, which can simplify the development process. Second, it is common that geoscience data has discontinuity in its hierarchy structure. The application of interface automata can prevent users from suffering automation surprises, and enhance user experience. Third, for supporting a variety of different data visualization and analysis, our design with interface automata could also make applications become extendable in that a new visualization function or a new data group could be easily added to an existing application, which reduces the overhead of maintenance significantly. We demonstrate the effectiveness of our framework using real-world applications.

  17. Structure of the automatic system for plasma equilibrium position control

    International Nuclear Information System (INIS)

    Gubarev, V.F.; Krivonos, Yu.G.; Samojlenko, Yu.I.; Snegur, A.A.

    1978-01-01

    Considered are the principles of construction of the automatic system for plasma filament equilibrium position control inside the discharge chamber for the installation of a tokamak type. The combined current control system in control winding is suggested. The most powerful subsystem creates current in the control winding according to the program calculated beforehand. This system provides plasma rough equilibrium along the ''big radius''. The subsystem performing the current change in small limits according to the principle of feed-back coupling is provided simultaneously. The stabilization of plasma position is achieved in the discharge chamber. The advantage of construction of such system is in decreasing of the automatic requlator power without lowering the requirements to the accuracy of equilibrium preservation. The subsystem of automatic control of plasma position over the vertical is put into the system. Such an approach to the construction of the automatic control system proves to be correct; it is based on the experience of application of similar devices for some existing thermonuclear plants

  18. A versatile Czochralski crystal growth system with automatic diameter control

    Science.gov (United States)

    Aggarwal, M. D.; Metzl, R.; Wang, W. S.; Choi, J.

    1995-07-01

    A versatile Czochralski crystal pulling system with automatic diameter control for the growth of nonlinear optical oxide crystals is discussed. Pure and doped bulk single crystals of bismuth silicon oxide (Bi12SiO20) have been successfully grown using this system. The system consists of a regular Czochralski type pulling system with provision for continuous weighing of the growing crystal to provide feedback for power control.

  19. Automatic system for evaluation of ionizing field

    International Nuclear Information System (INIS)

    Pimenta, N.L.; Calil, S.J.

    1992-01-01

    A three-dimensional cartesian manipulator for evaluating the ionizing field and able to position a ionization chamber in any point of the space is developed. The control system is made using a IBM microcomputer. The system aimed the study of isodose curves from ionizing sources, verifying the performance of radiotherapeutic equipment. (C.G.C.)

  20. Diagnosis - Using automatic test equipment and artificial intelligence expert systems

    Science.gov (United States)

    Ramsey, J. E., Jr.

    Three expert systems (ATEOPS, ATEFEXPERS, and ATEFATLAS), which were created to direct automatic test equipment (ATE), are reviewed. The purpose of the project was to develop an expert system to troubleshoot the converter-programmer power supply card for the F-15 aircraft and have that expert system direct the automatic test equipment. Each expert system uses a different knowledge base or inference engine, basing the testing on the circuit schematic, test requirements document, or ATLAS code. Implementing generalized modules allows the expert systems to be used for any different unit under test. Using converted ATLAS to LISP code allows the expert system to direct any ATE using ATLAS. The constraint propagated frame system allows for the expansion of control by creating the ATLAS code, checking the code for good software engineering techniques, directing the ATE, and changing the test sequence as needed (planning).