WorldWideScience

Sample records for point source database

  1. Geodetic Control Points - Multi-State Control Point Database

    Data.gov (United States)

    NSGIC State | GIS Inventory — The Multi-State Control Point Database (MCPD) is a database of geodetic and mapping control covering Idaho and Montana. The control were submitted by registered land...

  2. Mobile Source Observation Database (MSOD)

    Science.gov (United States)

    The Mobile Source Observation Database (MSOD) is a relational database developed by the Assessment and Standards Division (ASD) of the U.S. EPA Office of Transportation and Air Quality (formerly the Office of Mobile Sources).

  3. Mobile Source Observation Database (MSOD)

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Mobile Source Observation Database (MSOD) is a relational database being developed by the Assessment and Standards Division (ASD) of the US Environmental...

  4. Free software and open source databases

    Directory of Open Access Journals (Sweden)

    Napoleon Alexandru SIRITEANU

    2006-01-01

    Full Text Available The emergence of free/open source software -FS/OSS- enterprises seeks to push software development out of the academic stream into the commercial mainstream, and as a result, end-user applications such as open source database management systems (PostgreSQL, MySQL, Firebird are becoming more popular. Companies like Sybase, Oracle, Sun, IBM are increasingly implementing open source strategies and porting programs/applications into the Linux environment. Open source software is redefining the software industry in general and database development in particular.

  5. Power source roadmaps using bibliometrics and database tomography

    International Nuclear Information System (INIS)

    Kostoff, R.N.; Tshiteya, R.; Pfeil, K.M.; Humenik, J.A.; Karypis, G.

    2005-01-01

    Database Tomography (DT) is a textual database analysis system consisting of two major components: (1) algorithms for extracting multi-word phrase frequencies and phrase proximities (physical closeness of the multi-word technical phrases) from any type of large textual database, to augment (2) interpretative capabilities of the expert human analyst. DT was used to derive technical intelligence from a Power Sources database derived from the Science Citation Index. Phrase frequency analysis by the technical domain experts provided the pervasive technical themes of the Power Sources database, and the phrase proximity analysis provided the relationships among the pervasive technical themes. Bibliometric analysis of the Power Sources literature supplemented the DT results with author/journal/institution/country publication and citation data

  6. Pseudo-dynamic source modelling with 1-point and 2-point statistics of earthquake source parameters

    KAUST Repository

    Song, S. G.

    2013-12-24

    Ground motion prediction is an essential element in seismic hazard and risk analysis. Empirical ground motion prediction approaches have been widely used in the community, but efficient simulation-based ground motion prediction methods are needed to complement empirical approaches, especially in the regions with limited data constraints. Recently, dynamic rupture modelling has been successfully adopted in physics-based source and ground motion modelling, but it is still computationally demanding and many input parameters are not well constrained by observational data. Pseudo-dynamic source modelling keeps the form of kinematic modelling with its computational efficiency, but also tries to emulate the physics of source process. In this paper, we develop a statistical framework that governs the finite-fault rupture process with 1-point and 2-point statistics of source parameters in order to quantify the variability of finite source models for future scenario events. We test this method by extracting 1-point and 2-point statistics from dynamically derived source models and simulating a number of rupture scenarios, given target 1-point and 2-point statistics. We propose a new rupture model generator for stochastic source modelling with the covariance matrix constructed from target 2-point statistics, that is, auto- and cross-correlations. Our sensitivity analysis of near-source ground motions to 1-point and 2-point statistics of source parameters provides insights into relations between statistical rupture properties and ground motions. We observe that larger standard deviation and stronger correlation produce stronger peak ground motions in general. The proposed new source modelling approach will contribute to understanding the effect of earthquake source on near-source ground motion characteristics in a more quantitative and systematic way.

  7. DEIMOS – an Open Source Image Database

    Directory of Open Access Journals (Sweden)

    M. Blazek

    2011-12-01

    Full Text Available The DEIMOS (DatabasE of Images: Open Source is created as an open-source database of images and videos for testing, verification and comparing of various image and/or video processing techniques such as enhancing, compression and reconstruction. The main advantage of DEIMOS is its orientation to various application fields – multimedia, television, security, assistive technology, biomedicine, astronomy etc. The DEIMOS is/will be created gradually step-by-step based upon the contributions of team members. The paper is describing basic parameters of DEIMOS database including application examples.

  8. Photoacoustic Point Source

    International Nuclear Information System (INIS)

    Calasso, Irio G.; Craig, Walter; Diebold, Gerald J.

    2001-01-01

    We investigate the photoacoustic effect generated by heat deposition at a point in space in an inviscid fluid. Delta-function and long Gaussian optical pulses are used as sources in the wave equation for the displacement potential to determine the fluid motion. The linear sound-generation mechanism gives bipolar photoacoustic waves, whereas the nonlinear mechanism produces asymmetric tripolar waves. The salient features of the photoacoustic point source are that rapid heat deposition and nonlinear thermal expansion dominate the production of ultrasound

  9. Pseudo-dynamic source modelling with 1-point and 2-point statistics of earthquake source parameters

    KAUST Repository

    Song, S. G.; Dalguer, L. A.; Mai, Paul Martin

    2013-01-01

    statistical framework that governs the finite-fault rupture process with 1-point and 2-point statistics of source parameters in order to quantify the variability of finite source models for future scenario events. We test this method by extracting 1-point

  10. Calcareous Fens - Source Feature Points

    Data.gov (United States)

    Minnesota Department of Natural Resources — Pursuant to the provisions of Minnesota Statutes, section 103G.223, this database contains points that represent calcareous fens as defined in Minnesota Rules, part...

  11. Open Source Vulnerability Database Project

    Directory of Open Access Journals (Sweden)

    Jake Kouns

    2008-06-01

    Full Text Available This article introduces the Open Source Vulnerability Database (OSVDB project which manages a global collection of computer security vulnerabilities, available for free use by the information security community. This collection contains information on known security weaknesses in operating systems, software products, protocols, hardware devices, and other infrastructure elements of information technology. The OSVDB project is intended to be the centralized global open source vulnerability collection on the Internet.

  12. A New Global Open Source Marine Hydrocarbon Emission Site Database

    Science.gov (United States)

    Onyia, E., Jr.; Wood, W. T.; Barnard, A.; Dada, T.; Qazzaz, M.; Lee, T. R.; Herrera, E.; Sager, W.

    2017-12-01

    Hydrocarbon emission sites (e.g. seeps) discharge large volumes of fluids and gases into the oceans that are not only important for biogeochemical budgets, but also support abundant chemosynthetic communities. Documenting the locations of modern emissions is a first step towards understanding and monitoring how they affect the global state of the seafloor and oceans. Currently, no global open source (i.e. non-proprietry) detailed maps of emissions sites are available. As a solution, we have created a database that is housed within an Excel spreadsheet and use the latest versions of Earthpoint and Google Earth for position coordinate conversions and data mapping, respectively. To date, approximately 1,000 data points have been collected from referenceable sources across the globe, and we are continualy expanding the dataset. Due to the variety of spatial extents encountered, to identify each site we used two different methods: 1) point (x, y, z) locations for individual sites and; 2) delineation of areas where sites are clustered. Certain well-known areas, such as the Gulf of Mexico and the Mediterranean Sea, have a greater abundance of information; whereas significantly less information is available in other regions due to the absence of emission sites, lack of data, or because the existing data is proprietary. Although the geographical extent of the data is currently restricted to regions where the most data is publicly available, as the database matures, we expect to have more complete coverage of the world's oceans. This database is an information resource that consolidates and organizes the existing literature on hydrocarbons released into the marine environment, thereby providing a comprehensive reference for future work. We expect that the availability of seafloor hydrocarbon emission maps will benefit scientific understanding of hydrocarbon rich areas as well as potentially aiding hydrocarbon exploration and environmental impact assessements.

  13. Modeling the contribution of point sources and non-point sources to Thachin River water pollution.

    Science.gov (United States)

    Schaffner, Monika; Bader, Hans-Peter; Scheidegger, Ruth

    2009-08-15

    Major rivers in developing and emerging countries suffer increasingly of severe degradation of water quality. The current study uses a mathematical Material Flow Analysis (MMFA) as a complementary approach to address the degradation of river water quality due to nutrient pollution in the Thachin River Basin in Central Thailand. This paper gives an overview of the origins and flow paths of the various point- and non-point pollution sources in the Thachin River Basin (in terms of nitrogen and phosphorus) and quantifies their relative importance within the system. The key parameters influencing the main nutrient flows are determined and possible mitigation measures discussed. The results show that aquaculture (as a point source) and rice farming (as a non-point source) are the key nutrient sources in the Thachin River Basin. Other point sources such as pig farms, households and industries, which were previously cited as the most relevant pollution sources in terms of organic pollution, play less significant roles in comparison. This order of importance shifts when considering the model results for the provincial level. Crosschecks with secondary data and field studies confirm the plausibility of our simulations. Specific nutrient loads for the pollution sources are derived; these can be used for a first broad quantification of nutrient pollution in comparable river basins. Based on an identification of the sensitive model parameters, possible mitigation scenarios are determined and their potential to reduce the nutrient load evaluated. A comparison of simulated nutrient loads with measured nutrient concentrations shows that nutrient retention in the river system may be significant. Sedimentation in the slow flowing surface water network as well as nitrogen emission to the air from the warm oxygen deficient waters are certainly partly responsible, but also wetlands along the river banks could play an important role as nutrient sinks.

  14. Source splitting via the point source method

    International Nuclear Information System (INIS)

    Potthast, Roland; Fazi, Filippo M; Nelson, Philip A

    2010-01-01

    We introduce a new algorithm for source identification and field splitting based on the point source method (Potthast 1998 A point-source method for inverse acoustic and electromagnetic obstacle scattering problems IMA J. Appl. Math. 61 119–40, Potthast R 1996 A fast new method to solve inverse scattering problems Inverse Problems 12 731–42). The task is to separate the sound fields u j , j = 1, ..., n of n element of N sound sources supported in different bounded domains G 1 , ..., G n in R 3 from measurements of the field on some microphone array—mathematically speaking from the knowledge of the sum of the fields u = u 1 + ... + u n on some open subset Λ of a plane. The main idea of the scheme is to calculate filter functions g 1 ,…, g n , n element of N, to construct u l for l = 1, ..., n from u| Λ in the form u l (x) = ∫ Λ g l,x (y)u(y)ds(y), l=1,... n. (1) We will provide the complete mathematical theory for the field splitting via the point source method. In particular, we describe uniqueness, solvability of the problem and convergence and stability of the algorithm. In the second part we describe the practical realization of the splitting for real data measurements carried out at the Institute for Sound and Vibration Research at Southampton, UK. A practical demonstration of the original recording and the splitting results for real data is available online

  15. Point Pollution Sources Dimensioning

    Directory of Open Access Journals (Sweden)

    Georgeta CUCULEANU

    2011-06-01

    Full Text Available In this paper a method for determining the main physical characteristics of the point pollution sources is presented. It can be used to find the main physical characteristics of them. The main physical characteristics of these sources are top inside source diameter and physical height. The top inside source diameter is calculated from gas flow-rate. For reckoning the physical height of the source one takes into account the relation given by the proportionality factor, defined as ratio between the plume rise and physical height of the source. The plume rise depends on the gas exit velocity and gas temperature. That relation is necessary for diminishing the environmental pollution when the production capacity of the plant varies, in comparison with the nominal one.

  16. Research on point source simulating the γ-ray detection efficiencies of stander source

    International Nuclear Information System (INIS)

    Tian Zining; Jia Mingyan; Shen Maoquan; Yang Xiaoyan; Cheng Zhiwei

    2010-01-01

    For φ 75 mm x 25 mm sample, the full energy peak efficiencies on different heights of sample radius were obtained using the point sources, and the function parameters about the full energy peak efficiencies of point sources based on radius was fixed. The 59.54 keV γ-ray, 661.66 keV γ-ray, 1173.2 keV γ-ray, 1332.5 keV γ-ray detection efficiencies on different height of samples were obtained, based on the full energy peak efficiencies of point sources and its height, and the function parameters about the full energy peak efficiencies of surface sources based on sample height was fixed. The detection efficiency of (75 mm x 25 mm calibration source can be obtained by integrality, the detection efficiencies simulated by point sources are consistent with the results of stander source in 10%. Therefore, the calibration method of stander source can be substituted by the point source simulation method, and it tis feasible when there is no stander source.) (authors)

  17. Induced Temporal Signatures for Point-Source Detection

    International Nuclear Information System (INIS)

    Stephens, Daniel L.; Runkle, Robert C.; Carlson, Deborah K.; Peurrung, Anthony J.; Seifert, Allen; Wyatt, Cory R.

    2005-01-01

    Detection of radioactive point-sized sources is inherently divided into two regimes encompassing stationary and moving detectors. The two cases differ in their treatment of background radiation and its influence on detection sensitivity. In the stationary detector case the statistical fluctuation of the background determines the minimum detectable quantity. In the moving detector case the detector may be subjected to widely and irregularly varying background radiation, as a result of geographical and environmental variation. This significant systematic variation, in conjunction with the statistical variation of the background, requires a conservative threshold to be selected to yield the same false-positive rate as the stationary detection case. This results in lost detection sensitivity for real sources. This work focuses on a simple and practical modification of the detector geometry that increase point-source recognition via a distinctive temporal signature. A key part of this effort is the integrated development of both detector geometries that induce a highly distinctive signature for point sources and the development of statistical algorithms able to optimize detection of this signature amidst varying background. The identification of temporal signatures for point sources has been demonstrated and compared with the canonical method showing good results. This work demonstrates that temporal signatures are efficient at increasing point-source discrimination in a moving detector system

  18. Systematic Review: Impact of point sources on antibiotic-resistant bacteria in the natural environment.

    Science.gov (United States)

    Bueno, I; Williams-Nguyen, J; Hwang, H; Sargeant, J M; Nault, A J; Singer, R S

    2018-02-01

    Point sources such as wastewater treatment plants and agricultural facilities may have a role in the dissemination of antibiotic-resistant bacteria (ARB) and antibiotic resistance genes (ARG). To analyse the evidence for increases in ARB in the natural environment associated with these point sources of ARB and ARG, we conducted a systematic review. We evaluated 5,247 records retrieved through database searches, including both studies that ascertained ARG and ARB outcomes. All studies were subjected to a screening process to assess relevance to the question and methodology to address our review question. A risk of bias assessment was conducted upon the final pool of studies included in the review. This article summarizes the evidence only for those studies with ARB outcomes (n = 47). Thirty-five studies were at high (n = 11) or at unclear (n = 24) risk of bias in the estimation of source effects due to lack of information and/or failure to control for confounders. Statistical analysis was used in ten studies, of which one assessed the effect of multiple sources using modelling approaches; none reported effect measures. Most studies reported higher ARB prevalence or concentration downstream/near the source. However, this evidence was primarily descriptive and it could not be concluded that there is a clear impact of point sources on increases in ARB in the environment. To quantify increases in ARB in the environment due to specific point sources, there is a need for studies that stress study design, control of biases and analytical tools to provide effect measure estimates. © 2017 Blackwell Verlag GmbH.

  19. Γ-source Neutral Point Clamped Inverter

    DEFF Research Database (Denmark)

    Mo, Wei; Loh, Poh Chiang; Blaabjerg, Frede

    Transformer based Z-source inverters are recently proposed to achieve promising buck-boost capability. They have improved higher buck-boost capability, smaller size and less components count over Z-source inverters. On the other hand, neutral point clamped inverters have less switching stress...... and better output performance comparing with traditional two-level inverters. Integrating these two types of configurations can help neutral point inverters achieve enhanced votlage buck-boost capability....

  20. Implementation of a database for the management of radioactive sources

    International Nuclear Information System (INIS)

    MOHAMAD, M.

    2012-01-01

    In Madagascar, the application of nuclear technology continues to develop. In order to protect the human health and his environment against the harmful effects of the ionizing radiation, each user of radioactive sources has to implement a program of nuclear security and safety and to declare their sources at Regulatory Authority. This Authority must have access to all the informations relating to all the sources and their uses. This work is based on the elaboration of a software using python as programming language and SQlite as database. It makes possible to computerize the radioactive sources management.This application unifies the various existing databases and centralizes the activities of the radioactive sources management.The objective is to follow the movement of each source in the Malagasy territory in order to avoid the risks related on the use of the radioactive sources and the illicit traffic. [fr

  1. Point-source inversion techniques

    Science.gov (United States)

    Langston, Charles A.; Barker, Jeffrey S.; Pavlin, Gregory B.

    1982-11-01

    A variety of approaches for obtaining source parameters from waveform data using moment-tensor or dislocation point source models have been investigated and applied to long-period body and surface waves from several earthquakes. Generalized inversion techniques have been applied to data for long-period teleseismic body waves to obtain the orientation, time function and depth of the 1978 Thessaloniki, Greece, event, of the 1971 San Fernando event, and of several events associated with the 1963 induced seismicity sequence at Kariba, Africa. The generalized inversion technique and a systematic grid testing technique have also been used to place meaningful constraints on mechanisms determined from very sparse data sets; a single station with high-quality three-component waveform data is often sufficient to discriminate faulting type (e.g., strike-slip, etc.). Sparse data sets for several recent California earthquakes, for a small regional event associated with the Koyna, India, reservoir, and for several events at the Kariba reservoir have been investigated in this way. Although linearized inversion techniques using the moment-tensor model are often robust, even for sparse data sets, there are instances where the simplifying assumption of a single point source is inadequate to model the data successfully. Numerical experiments utilizing synthetic data and actual data for the 1971 San Fernando earthquake graphically demonstrate that severe problems may be encountered if source finiteness effects are ignored. These techniques are generally applicable to on-line processing of high-quality digital data, but source complexity and inadequacy of the assumed Green's functions are major problems which are yet to be fully addressed.

  2. Oak Ridge Reservation Environmental Protection Rad Neshaps Radionuclide Inventory Web Database and Rad Neshaps Source and Dose Database.

    Science.gov (United States)

    Scofield, Patricia A; Smith, Linda L; Johnson, David N

    2017-07-01

    The U.S. Environmental Protection Agency promulgated national emission standards for emissions of radionuclides other than radon from US Department of Energy facilities in Chapter 40 of the Code of Federal Regulations (CFR) 61, Subpart H. This regulatory standard limits the annual effective dose that any member of the public can receive from Department of Energy facilities to 0.1 mSv. As defined in the preamble of the final rule, all of the facilities on the Oak Ridge Reservation, i.e., the Y-12 National Security Complex, Oak Ridge National Laboratory, East Tennessee Technology Park, and any other U.S. Department of Energy operations on Oak Ridge Reservation, combined, must meet the annual dose limit of 0.1 mSv. At Oak Ridge National Laboratory, there are monitored sources and numerous unmonitored sources. To maintain radiological source and inventory information for these unmonitored sources, e.g., laboratory hoods, equipment exhausts, and room exhausts not currently venting to monitored stacks on the Oak Ridge National Laboratory campus, the Environmental Protection Rad NESHAPs Inventory Web Database was developed. This database is updated annually and is used to compile emissions data for the annual Radionuclide National Emission Standards for Hazardous Air Pollutants (Rad NESHAPs) report required by 40 CFR 61.94. It also provides supporting documentation for facility compliance audits. In addition, a Rad NESHAPs source and dose database was developed to import the source and dose summary data from Clean Air Act Assessment Package-1988 computer model files. This database provides Oak Ridge Reservation and facility-specific source inventory; doses associated with each source and facility; and total doses for the Oak Ridge Reservation dose.

  3. UHE point source survey at Cygnus experiment

    International Nuclear Information System (INIS)

    Lu, X.; Yodh, G.B.; Alexandreas, D.E.; Allen, R.C.; Berley, D.; Biller, S.D.; Burman, R.L.; Cady, R.; Chang, C.Y.; Dingus, B.L.; Dion, G.M.; Ellsworth, R.W.; Gilra, M.K.; Goodman, J.A.; Haines, T.J.; Hoffman, C.M.; Kwok, P.; Lloyd-Evans, J.; Nagle, D.E.; Potter, M.E.; Sandberg, V.D.; Stark, M.J.; Talaga, R.L.; Vishwanath, P.R.; Zhang, W.

    1991-01-01

    A new method of searching for UHE point source has been developed. With a data sample of 150 million events, we have surveyed the sky for point sources over 3314 locations (1.4 degree <δ<70.4 degree). It was found that their distribution is consistent with a random fluctuation. In addition, fifty two known potential sources, including pulsars and binary x-ray sources, were studied. The source with the largest positive excess is the Crab Nebula. An excess of 2.5 sigma above the background is observed in a bin of 2.3 degree by 2.5 degree in declination and right ascension respectively

  4. Point source reconstruction principle of linear inverse problems

    International Nuclear Information System (INIS)

    Terazono, Yasushi; Matani, Ayumu; Fujimaki, Norio; Murata, Tsutomu

    2010-01-01

    Exact point source reconstruction for underdetermined linear inverse problems with a block-wise structure was studied. In a block-wise problem, elements of a source vector are partitioned into blocks. Accordingly, a leadfield matrix, which represents the forward observation process, is also partitioned into blocks. A point source is a source having only one nonzero block. An example of such a problem is current distribution estimation in electroencephalography and magnetoencephalography, where a source vector represents a vector field and a point source represents a single current dipole. In this study, the block-wise norm, a block-wise extension of the l p -norm, was defined as the family of cost functions of the inverse method. The main result is that a set of three conditions was found to be necessary and sufficient for block-wise norm minimization to ensure exact point source reconstruction for any leadfield matrix that admit such reconstruction. The block-wise norm that satisfies the conditions is the sum of the cost of all the observations of source blocks, or in other words, the block-wisely extended leadfield-weighted l 1 -norm. Additional results are that minimization of such a norm always provides block-wisely sparse solutions and that its solutions form cones in source space

  5. X pinch a point x-ray source

    International Nuclear Information System (INIS)

    Garg, A.B.; Rout, R.K.; Shyam, A.; Srinivasan, M.

    1993-01-01

    X ray emission from an X pinch, a point x-ray source has been studied using a pin-hole camera by a 30 kV, 7.2 μ F capacitor bank. The wires of different material like W, Mo, Cu, S.S.(stainless steel) and Ti were used. Molybdenum pinch gives the most intense x-rays and stainless steel gives the minimum intensity x-rays for same bank energy (∼ 3.2 kJ). Point x-ray source of size (≤ 0.5 mm) was observed using pin hole camera. The size of the source is limited by the size of the pin hole camera. The peak current in the load is approximately 150 kA. The point x-ray source could be useful in many fields like micro lithography, medicine and to study the basic physics of high Z plasmas. (author). 4 refs., 3 figs

  6. DABAM: an open-source database of X-ray mirrors metrology

    Energy Technology Data Exchange (ETDEWEB)

    Sanchez del Rio, Manuel, E-mail: srio@esrf.eu [ESRF - The European Synchrotron, 71 Avenue des Martyrs, 38000 Grenoble (France); Bianchi, Davide [AC2T Research GmbH, Viktro-Kaplan-Strasse 2-C, 2700 Wiener Neustadt (Austria); Cocco, Daniele [SLAC National Accelerator Laboratory, 2575 Sand Hill Road, Menlo Park, CA 94025 (United States); Glass, Mark [ESRF - The European Synchrotron, 71 Avenue des Martyrs, 38000 Grenoble (France); Idir, Mourad [NSLS II, Brookhaven National Laboratory, Upton, NY 11973-5000 (United States); Metz, Jim [InSync Inc., 2511C Broadbent Parkway, Albuquerque, NM 87107 (United States); Raimondi, Lorenzo; Rebuffi, Luca [Elettra-Sincrotrone Trieste SCpA, Basovizza (TS) (Italy); Reininger, Ruben; Shi, Xianbo [Advanced Photon Source, Argonne National Laboratory, Argonne, IL 60439 (United States); Siewert, Frank [BESSY II, Helmholtz Zentrum Berlin, Institute for Nanometre Optics and Technology, Albert-Einstein-Strasse 15, 12489 Berlin (Germany); Spielmann-Jaeggi, Sibylle [Swiss Light Source at Paul Scherrer Institut, CH-5232 Villigen PSI (Switzerland); Takacs, Peter [Instrumentation Division, Brookhaven National Laboratory, Upton, NY 11973-5000 (United States); Tomasset, Muriel [Synchrotron Soleil (France); Tonnessen, Tom [InSync Inc., 2511C Broadbent Parkway, Albuquerque, NM 87107 (United States); Vivo, Amparo [ESRF - The European Synchrotron, 71 Avenue des Martyrs, 38000 Grenoble (France); Yashchuk, Valeriy [Advanced Light Source, Lawrence Berkeley National Laboratory, MS 15-R0317, 1 Cyclotron Road, Berkeley, CA 94720-8199 (United States)

    2016-04-20

    DABAM, an open-source database of X-ray mirrors metrology to be used with ray-tracing and wave-propagation codes for simulating the effect of the surface errors on the performance of a synchrotron radiation beamline. An open-source database containing metrology data for X-ray mirrors is presented. It makes available metrology data (mirror heights and slopes profiles) that can be used with simulation tools for calculating the effects of optical surface errors in the performances of an optical instrument, such as a synchrotron beamline. A typical case is the degradation of the intensity profile at the focal position in a beamline due to mirror surface errors. This database for metrology (DABAM) aims to provide to the users of simulation tools the data of real mirrors. The data included in the database are described in this paper, with details of how the mirror parameters are stored. An accompanying software is provided to allow simple access and processing of these data, calculate the most usual statistical parameters, and also include the option of creating input files for most used simulation codes. Some optics simulations are presented and discussed to illustrate the real use of the profiles from the database.

  7. Unidentified point sources in the IRAS minisurvey

    Science.gov (United States)

    Houck, J. R.; Soifer, B. T.; Neugebauer, G.; Beichman, C. A.; Aumann, H. H.; Clegg, P. E.; Gillett, F. C.; Habing, H. J.; Hauser, M. G.; Low, F. J.

    1984-01-01

    Nine bright, point-like 60 micron sources have been selected from the sample of 8709 sources in the IRAS minisurvey. These sources have no counterparts in a variety of catalogs of nonstellar objects. Four objects have no visible counterparts, while five have faint stellar objects visible in the error ellipse. These sources do not resemble objects previously known to be bright infrared sources.

  8. The Potential for Electrofuels Production in Sweden Utilizing Fossil and Biogenic CO2 Point Sources

    International Nuclear Information System (INIS)

    Hansson, Julia; Hackl, Roman; Taljegard, Maria; Brynolf, Selma; Grahn, Maria

    2017-01-01

    This paper maps, categorizes, and quantifies all major point sources of carbon dioxide (CO 2 ) emissions from industrial and combustion processes in Sweden. The paper also estimates the Swedish technical potential for electrofuels (power-to-gas/fuels) based on carbon capture and utilization. With our bottom-up approach using European databases, we find that Sweden emits approximately 50 million metric tons of CO 2 per year from different types of point sources, with 65% (or about 32 million tons) from biogenic sources. The major sources are the pulp and paper industry (46%), heat and power production (23%), and waste treatment and incineration (8%). Most of the CO 2 is emitted at low concentrations (<15%) from sources in the southern part of Sweden where power demand generally exceeds in-region supply. The potentially recoverable emissions from all the included point sources amount to 45 million tons. If all the recoverable CO 2 were used to produce electrofuels, the yield would correspond to 2–3 times the current Swedish demand for transportation fuels. The electricity required would correspond to about 3 times the current Swedish electricity supply. The current relatively few emission sources with high concentrations of CO 2 (>90%, biofuel operations) would yield electrofuels corresponding to approximately 2% of the current demand for transportation fuels (corresponding to 1.5–2 TWh/year). In a 2030 scenario with large-scale biofuels operations based on lignocellulosic feedstocks, the potential for electrofuels production from high-concentration sources increases to 8–11 TWh/year. Finally, renewable electricity and production costs, rather than CO 2 supply, limit the potential for production of electrofuels in Sweden.

  9. 2011 Radioactive Materials Usage Survey for Unmonitored Point Sources

    Energy Technology Data Exchange (ETDEWEB)

    Sturgeon, Richard W. [Los Alamos National Laboratory

    2012-06-27

    This report provides the results of the 2011 Radioactive Materials Usage Survey for Unmonitored Point Sources (RMUS), which was updated by the Environmental Protection (ENV) Division's Environmental Stewardship (ES) at Los Alamos National Laboratory (LANL). ES classifies LANL emission sources into one of four Tiers, based on the potential effective dose equivalent (PEDE) calculated for each point source. Detailed descriptions of these tiers are provided in Section 3. The usage survey is conducted annually; in odd-numbered years the survey addresses all monitored and unmonitored point sources and in even-numbered years it addresses all Tier III and various selected other sources. This graded approach was designed to ensure that the appropriate emphasis is placed on point sources that have higher potential emissions to the environment. For calendar year (CY) 2011, ES has divided the usage survey into two distinct reports, one covering the monitored point sources (to be completed later this year) and this report covering all unmonitored point sources. This usage survey includes the following release points: (1) all unmonitored sources identified in the 2010 usage survey, (2) any new release points identified through the new project review (NPR) process, and (3) other release points as designated by the Rad-NESHAP Team Leader. Data for all unmonitored point sources at LANL is stored in the survey files at ES. LANL uses this survey data to help demonstrate compliance with Clean Air Act radioactive air emissions regulations (40 CFR 61, Subpart H). The remainder of this introduction provides a brief description of the information contained in each section. Section 2 of this report describes the methods that were employed for gathering usage survey data and for calculating usage, emissions, and dose for these point sources. It also references the appropriate ES procedures for further information. Section 3 describes the RMUS and explains how the survey results are

  10. [A landscape ecological approach for urban non-point source pollution control].

    Science.gov (United States)

    Guo, Qinghai; Ma, Keming; Zhao, Jingzhu; Yang, Liu; Yin, Chengqing

    2005-05-01

    Urban non-point source pollution is a new problem appeared with the speeding development of urbanization. The particularity of urban land use and the increase of impervious surface area make urban non-point source pollution differ from agricultural non-point source pollution, and more difficult to control. Best Management Practices (BMPs) are the effective practices commonly applied in controlling urban non-point source pollution, mainly adopting local repairing practices to control the pollutants in surface runoff. Because of the close relationship between urban land use patterns and non-point source pollution, it would be rational to combine the landscape ecological planning with local BMPs to control the urban non-point source pollution, which needs, firstly, analyzing and evaluating the influence of landscape structure on water-bodies, pollution sources and pollutant removal processes to define the relationships between landscape spatial pattern and non-point source pollution and to decide the key polluted fields, and secondly, adjusting inherent landscape structures or/and joining new landscape factors to form new landscape pattern, and combining landscape planning and management through applying BMPs into planning to improve urban landscape heterogeneity and to control urban non-point source pollution.

  11. On the point-source approximation of earthquake dynamics

    Directory of Open Access Journals (Sweden)

    Andrea Bizzarri

    2014-06-01

    Full Text Available The focus on the present study is on the point-source approximation of a seismic source. First, we compare the synthetic motions on the free surface resulting from different analytical evolutions of the seismic source (the Gabor signal (G, the Bouchon ramp (B, the Cotton and Campillo ramp (CC, the Yoffe function (Y and the Liu and Archuleta function (LA. Our numerical experiments indicate that the CC and the Y functions produce synthetics with larger oscillations and correspondingly they have a higher frequency content. Moreover, the CC and the Y functions tend to produce higher peaks in the ground velocity (roughly of a factor of two. We have also found that the falloff at high frequencies is quite different: it roughly follows ω−2 in the case of G and LA functions, it decays more faster than ω−2 for the B function, while it is slow than ω−1 for both the CC and the Y solutions. Then we perform a comparison of seismic waves resulting from 3-D extended ruptures (both supershear and subshear obeying to different governing laws against those from a single point-source having the same features. It is shown that the point-source models tend to overestimate the ground motions and that they completely miss the Mach fronts emerging from the supershear transition process. When we compare the extended fault solutions against a multiple point-sources model the agreement becomes more significant, although relevant discrepancies still persist. Our results confirm that, and more importantly quantify how, the point-source approximation is unable to adequately describe the radiation emitted during a real world earthquake, even in the most idealized case of planar fault with homogeneous properties and embedded in a homogeneous, perfectly elastic medium.

  12. The Potential for Electrofuels Production in Sweden Utilizing Fossil and Biogenic CO{sub 2} Point Sources

    Energy Technology Data Exchange (ETDEWEB)

    Hansson, Julia, E-mail: julia.hansson@ivl.se [Climate and Sustainable Cities, IVL Swedish Environmental Research Institute, Stockholm (Sweden); Division of Physical Resource Theory, Department of Energy and Environment, Chalmers University of Technology, Göteborg (Sweden); Hackl, Roman [Climate and Sustainable Cities, IVL Swedish Environmental Research Institute, Stockholm (Sweden); Taljegard, Maria [Division of Energy Technology, Department of Energy and Environment, Chalmers University of Technology, Göteborg (Sweden); Brynolf, Selma; Grahn, Maria [Division of Physical Resource Theory, Department of Energy and Environment, Chalmers University of Technology, Göteborg (Sweden)

    2017-03-13

    This paper maps, categorizes, and quantifies all major point sources of carbon dioxide (CO{sub 2}) emissions from industrial and combustion processes in Sweden. The paper also estimates the Swedish technical potential for electrofuels (power-to-gas/fuels) based on carbon capture and utilization. With our bottom-up approach using European databases, we find that Sweden emits approximately 50 million metric tons of CO{sub 2} per year from different types of point sources, with 65% (or about 32 million tons) from biogenic sources. The major sources are the pulp and paper industry (46%), heat and power production (23%), and waste treatment and incineration (8%). Most of the CO{sub 2} is emitted at low concentrations (<15%) from sources in the southern part of Sweden where power demand generally exceeds in-region supply. The potentially recoverable emissions from all the included point sources amount to 45 million tons. If all the recoverable CO{sub 2} were used to produce electrofuels, the yield would correspond to 2–3 times the current Swedish demand for transportation fuels. The electricity required would correspond to about 3 times the current Swedish electricity supply. The current relatively few emission sources with high concentrations of CO{sub 2} (>90%, biofuel operations) would yield electrofuels corresponding to approximately 2% of the current demand for transportation fuels (corresponding to 1.5–2 TWh/year). In a 2030 scenario with large-scale biofuels operations based on lignocellulosic feedstocks, the potential for electrofuels production from high-concentration sources increases to 8–11 TWh/year. Finally, renewable electricity and production costs, rather than CO{sub 2} supply, limit the potential for production of electrofuels in Sweden.

  13. Rainfall Deduction Method for Estimating Non-Point Source Pollution Load for Watershed

    OpenAIRE

    Cai, Ming; Li, Huai-en; KAWAKAMI, Yoji

    2004-01-01

    The water pollution can be divided into point source pollution (PSP) and non-point source pollution (NSP). Since the point source pollution has been controlled, the non-point source pollution is becoming the main pollution source. The prediction of NSP load is being increasingly important in water pollution controlling and planning in watershed. Considering the monitoring data shortage of NPS in China, a practical estimation method of non-point source pollution load --- rainfall deduction met...

  14. Comparison of open source database systems(characteristics, limits of usage)

    OpenAIRE

    Husárik, Braňko

    2008-01-01

    The goal of this work is to compare some chosen open source database systems (Ingres, PostgreSQL, Firebird, Mysql). First part of work is focused on history and present situation of companies which are developing these products. Second part contains the comparision of certain group of specific features and limits. The benchmark of some operations is its own part. Possibilities of usage of mentioned database systems are summarized at the end of work.

  15. Very Luminous X-ray Point Sources in Starburst Galaxies

    Science.gov (United States)

    Colbert, E.; Heckman, T.; Ptak, A.; Weaver, K. A.; Strickland, D.

    Extranuclear X-ray point sources in external galaxies with luminosities above 1039.0 erg/s are quite common in elliptical, disk and dwarf galaxies, with an average of ~ 0.5 and dwarf galaxies, with an average of ~0.5 sources per galaxy. These objects may be a new class of object, perhaps accreting intermediate-mass black holes, or beamed stellar mass black hole binaries. Starburst galaxies tend to have a larger number of these intermediate-luminosity X-ray objects (IXOs), as well as a large number of lower-luminosity (1037 - 1039 erg/s) point sources. These point sources dominate the total hard X-ray emission in starburst galaxies. We present a review of both types of objects and discuss possible schemes for their formation.

  16. Healthcare Databases in Thailand and Japan: Potential Sources for Health Technology Assessment Research.

    Directory of Open Access Journals (Sweden)

    Surasak Saokaew

    Full Text Available Health technology assessment (HTA has been continuously used for value-based healthcare decisions over the last decade. Healthcare databases represent an important source of information for HTA, which has seen a surge in use in Western countries. Although HTA agencies have been established in Asia-Pacific region, application and understanding of healthcare databases for HTA is rather limited. Thus, we reviewed existing databases to assess their potential for HTA in Thailand where HTA has been used officially and Japan where HTA is going to be officially introduced.Existing healthcare databases in Thailand and Japan were compiled and reviewed. Databases' characteristics e.g. name of database, host, scope/objective, time/sample size, design, data collection method, population/sample, and variables were described. Databases were assessed for its potential HTA use in terms of safety/efficacy/effectiveness, social/ethical, organization/professional, economic, and epidemiological domains. Request route for each database was also provided.Forty databases- 20 from Thailand and 20 from Japan-were included. These comprised of national censuses, surveys, registries, administrative data, and claimed databases. All databases were potentially used for epidemiological studies. In addition, data on mortality, morbidity, disability, adverse events, quality of life, service/technology utilization, length of stay, and economics were also found in some databases. However, access to patient-level data was limited since information about the databases was not available on public sources.Our findings have shown that existing databases provided valuable information for HTA research with limitation on accessibility. Mutual dialogue on healthcare database development and usage for HTA among Asia-Pacific region is needed.

  17. Healthcare Databases in Thailand and Japan: Potential Sources for Health Technology Assessment Research.

    Science.gov (United States)

    Saokaew, Surasak; Sugimoto, Takashi; Kamae, Isao; Pratoomsoot, Chayanin; Chaiyakunapruk, Nathorn

    2015-01-01

    Health technology assessment (HTA) has been continuously used for value-based healthcare decisions over the last decade. Healthcare databases represent an important source of information for HTA, which has seen a surge in use in Western countries. Although HTA agencies have been established in Asia-Pacific region, application and understanding of healthcare databases for HTA is rather limited. Thus, we reviewed existing databases to assess their potential for HTA in Thailand where HTA has been used officially and Japan where HTA is going to be officially introduced. Existing healthcare databases in Thailand and Japan were compiled and reviewed. Databases' characteristics e.g. name of database, host, scope/objective, time/sample size, design, data collection method, population/sample, and variables were described. Databases were assessed for its potential HTA use in terms of safety/efficacy/effectiveness, social/ethical, organization/professional, economic, and epidemiological domains. Request route for each database was also provided. Forty databases- 20 from Thailand and 20 from Japan-were included. These comprised of national censuses, surveys, registries, administrative data, and claimed databases. All databases were potentially used for epidemiological studies. In addition, data on mortality, morbidity, disability, adverse events, quality of life, service/technology utilization, length of stay, and economics were also found in some databases. However, access to patient-level data was limited since information about the databases was not available on public sources. Our findings have shown that existing databases provided valuable information for HTA research with limitation on accessibility. Mutual dialogue on healthcare database development and usage for HTA among Asia-Pacific region is needed.

  18. The Protein Identifier Cross-Referencing (PICR service: reconciling protein identifiers across multiple source databases

    Directory of Open Access Journals (Sweden)

    Leinonen Rasko

    2007-10-01

    Full Text Available Abstract Background Each major protein database uses its own conventions when assigning protein identifiers. Resolving the various, potentially unstable, identifiers that refer to identical proteins is a major challenge. This is a common problem when attempting to unify datasets that have been annotated with proteins from multiple data sources or querying data providers with one flavour of protein identifiers when the source database uses another. Partial solutions for protein identifier mapping exist but they are limited to specific species or techniques and to a very small number of databases. As a result, we have not found a solution that is generic enough and broad enough in mapping scope to suit our needs. Results We have created the Protein Identifier Cross-Reference (PICR service, a web application that provides interactive and programmatic (SOAP and REST access to a mapping algorithm that uses the UniProt Archive (UniParc as a data warehouse to offer protein cross-references based on 100% sequence identity to proteins from over 70 distinct source databases loaded into UniParc. Mappings can be limited by source database, taxonomic ID and activity status in the source database. Users can copy/paste or upload files containing protein identifiers or sequences in FASTA format to obtain mappings using the interactive interface. Search results can be viewed in simple or detailed HTML tables or downloaded as comma-separated values (CSV or Microsoft Excel (XLS files suitable for use in a local database or a spreadsheet. Alternatively, a SOAP interface is available to integrate PICR functionality in other applications, as is a lightweight REST interface. Conclusion We offer a publicly available service that can interactively map protein identifiers and protein sequences to the majority of commonly used protein databases. Programmatic access is available through a standards-compliant SOAP interface or a lightweight REST interface. The PICR

  19. Family of Quantum Sources for Improving Near Field Accuracy in Transducer Modeling by the Distributed Point Source Method

    Directory of Open Access Journals (Sweden)

    Dominique Placko

    2016-10-01

    Full Text Available The distributed point source method, or DPSM, developed in the last decade has been used for solving various engineering problems—such as elastic and electromagnetic wave propagation, electrostatic, and fluid flow problems. Based on a semi-analytical formulation, the DPSM solution is generally built by superimposing the point source solutions or Green’s functions. However, the DPSM solution can be also obtained by superimposing elemental solutions of volume sources having some source density called the equivalent source density (ESD. In earlier works mostly point sources were used. In this paper the DPSM formulation is modified to introduce a new kind of ESD, replacing the classical single point source by a family of point sources that are referred to as quantum sources. The proposed formulation with these quantum sources do not change the dimension of the global matrix to be inverted to solve the problem when compared with the classical point source-based DPSM formulation. To assess the performance of this new formulation, the ultrasonic field generated by a circular planer transducer was compared with the classical DPSM formulation and analytical solution. The results show a significant improvement in the near field computation.

  20. Calculation of dose for β point and sphere sources in soft tissue

    International Nuclear Information System (INIS)

    Sun Fuyin; Yuan Shuyu; Tan Jian

    1999-01-01

    Objective: To compare the results of the distribution of dose rate calculated by three typical methods for point source and sphere source of β nuclide. Methods: Calculating and comparing the distributions of dose rate from 32 P β point and sphere sources in soft tissue calculated by the three methods published in references, [1]. [2] and [3], respectively. Results: For the point source of 3.7 x 10 7 Bq (1mCi), the variations of the calculation results of the three formulas are within 10% if r≤0.35 g/cm 2 , r being the distance from source, and larger than 10% if r > 0.35 g/cm 2 . For the sphere source whose volume is 50 μl and activity is 3.7 x 10 7 Bq(1 mCi), the variations are within 10% if z≤0.15 g/cm 2 , z being the distance from the surface of the sphere source to a point outside the sphere. Conclusion: The agreement of the distributions of the dose rate calculated by the three methods mentioned above for point and sphere β source are good if the distances from point source or the surface of sphere source to the points observed are small, and poor if they are large

  1. The NASA Goddard Group's Source Monitoring Database and Program

    Science.gov (United States)

    Gipson, John; Le Bail, Karine; Ma, Chopo

    2014-12-01

    Beginning in 2003, the Goddard VLBI group developed a program to purposefully monitor when sources were observed and to increase the observations of ``under-observed'' sources. The heart of the program consists of a MySQL database that keeps track of, on a session-by-session basis: the number of observations that are scheduled for a source, the number of observations that are successfully correlated, and the number of observations that are used in a session. In addition, there is a table that contains the target number of successful sessions over the last twelve months. Initially this table just contained two categories. Sources in the geodetic catalog had a target of 12 sessions/year; the remaining ICRF-1 defining sources had a target of two sessions/year. All other sources did not have a specific target. As the program evolved, different kinds of sources with different observing targets were added. During the scheduling process, the scheduler has the option of automatically selecting N sources which have not met their target. We discuss the history and present some results of this successful program.

  2. X-ray Point Source Populations in Spiral and Elliptical Galaxies

    Science.gov (United States)

    Colbert, E.; Heckman, T.; Weaver, K.; Strickland, D.

    2002-01-01

    The hard-X-ray luminosity of non-active galaxies has been known to be fairly well correlated with the total blue luminosity since the days of the Einstein satellite. However, the origin of this hard component was not well understood. Some possibilities that were considered included X-ray binaries, extended upscattered far-infrared light via the inverse-Compton process, extended hot 107 K gas (especially in ellipitical galaxies), or even an active nucleus. Chandra images of normal, elliptical and starburst galaxies now show that a significant amount of the total hard X-ray emission comes from individual point sources. We present here spatial and spectral analyses of the point sources in a small sample of Chandra obervations of starburst galaxies, and compare with Chandra point source analyses from comparison galaxies (elliptical, Seyfert and normal galaxies). We discuss possible relationships between the number and total hard luminosity of the X-ray point sources and various measures of the galaxy star formation rate, and discuss possible options for the numerous compact sources that are observed.

  3. [Nitrogen non-point source pollution identification based on ArcSWAT in Changle River].

    Science.gov (United States)

    Deng, Ou-Ping; Sun, Si-Yang; Lü, Jun

    2013-04-01

    The ArcSWAT (Soil and Water Assessment Tool) model was adopted for Non-point source (NPS) nitrogen pollution modeling and nitrogen source apportionment for the Changle River watershed, a typical agricultural watershed in Southeast China. Water quality and hydrological parameters were monitored, and the watershed natural conditions (including soil, climate, land use, etc) and pollution sources information were also investigated and collected for SWAT database. The ArcSWAT model was established in the Changle River after the calibrating and validating procedures of the model parameters. Based on the validated SWAT model, the contributions of different nitrogen sources to river TN loading were quantified, and spatial-temporal distributions of NPS nitrogen export to rivers were addressed. The results showed that in the Changle River watershed, Nitrogen fertilizer, nitrogen air deposition and nitrogen soil pool were the prominent pollution sources, which contributed 35%, 32% and 25% to the river TN loading, respectively. There were spatial-temporal variations in the critical sources for NPS TN export to the river. Natural sources, such as soil nitrogen pool and atmospheric nitrogen deposition, should be targeted as the critical sources for river TN pollution during the rainy seasons. Chemical nitrogen fertilizer application should be targeted as the critical sources for river TN pollution during the crop growing season. Chemical nitrogen fertilizer application, soil nitrogen pool and atmospheric nitrogen deposition were the main sources for TN exported from the garden plot, forest and residential land, respectively. However, they were the main sources for TN exported both from the upland and paddy field. These results revealed that NPS pollution controlling rules should focus on the spatio-temporal distribution of NPS pollution sources.

  4. Summary of Adsorption/Desorption Experiments for the European Database on Indoor Air Pollution Sources in Buildings

    DEFF Research Database (Denmark)

    Kjær, Ulla Dorte; Tirkkonen, T.

    1996-01-01

    Experimental data for adsorption/desorption in building materials. Contribution to the European Database on Indoor Air Pollution Sources in buildings.......Experimental data for adsorption/desorption in building materials. Contribution to the European Database on Indoor Air Pollution Sources in buildings....

  5. Trans-Z-source Neutral Point Clamped inverter

    DEFF Research Database (Denmark)

    Mo, W.; Loh, P. C.; Li, D.

    2012-01-01

    Transformer based Z-source (trans-Z-source) inverters are recently proposed by extending the traditional Z-source inverter with higher buck-boost capability as well as reducing the passive components at the same time. Multi-Level Z-source inverters are single-stage topological solutions used...... for buck-boost energy conversion with all the favourable advantages of multi-level switching retained. This paper presents three-level trans-Z-source Neutral Point Clamped (NPC) inverter topology, which achieves both the advantages of trans-Z-source and three-level NPC inverter configuration. With proper...... modulation scheme, the three-level trans-Z-source inverter can function with minimum of six device commutations per half carrier cycle (same as the traditional buck NPC inverter), while maintaining to produce the designed volt-sec average and inductive voltage boosting at ac output terminals. The designed...

  6. Determination of disintegration rates of a 60Co point source and volume sources by the sum-peak method

    International Nuclear Information System (INIS)

    Kawano, Takao; Ebihara, Hiroshi

    1990-01-01

    The disintegration rates of 60 Co as a point source (<2 mm in diameter on a thin plastic disc) and volume sources (10-100 mL solutions in a polyethylene bottle) are determined by the sum-peak method. The sum-peak formula gives the exact disintegration rate for the point source at different positions from the detector. However, increasing the volume of the solution results in enlarged deviations from the true disintegration rate. Extended sources must be treated as an amalgam of many point sources. (author)

  7. Analysis of point source size on measurement accuracy of lateral point-spread function of confocal Raman microscopy

    Science.gov (United States)

    Fu, Shihang; Zhang, Li; Hu, Yao; Ding, Xiang

    2018-01-01

    Confocal Raman Microscopy (CRM) has matured to become one of the most powerful instruments in analytical science because of its molecular sensitivity and high spatial resolution. Compared with conventional Raman Microscopy, CRM can perform three dimensions mapping of tiny samples and has the advantage of high spatial resolution thanking to the unique pinhole. With the wide application of the instrument, there is a growing requirement for the evaluation of the imaging performance of the system. Point-spread function (PSF) is an important approach to the evaluation of imaging capability of an optical instrument. Among a variety of measurement methods of PSF, the point source method has been widely used because it is easy to operate and the measurement results are approximate to the true PSF. In the point source method, the point source size has a significant impact on the final measurement accuracy. In this paper, the influence of the point source sizes on the measurement accuracy of PSF is analyzed and verified experimentally. A theoretical model of the lateral PSF for CRM is established and the effect of point source size on full-width at half maximum of lateral PSF is simulated. For long-term preservation and measurement convenience, PSF measurement phantom using polydimethylsiloxane resin, doped with different sizes of polystyrene microspheres is designed. The PSF of CRM with different sizes of microspheres are measured and the results are compared with the simulation results. The results provide a guide for measuring the PSF of the CRM.

  8. Concept for Risk-based Prioritisation of Point Sources

    DEFF Research Database (Denmark)

    Overheu, N.D.; Troldborg, Mads; Tuxen, N.

    2010-01-01

    estimates on a local scale from all the sources, and 3D catchment-scale fate and transport modelling. It handles point sources at various knowledge levels and accounts for uncertainties. The tool estimates the impacts on the water supply in the catchment and provides an overall prioritisation of the sites...

  9. Pan European Phenological database (PEP725): a single point of access for European data

    Science.gov (United States)

    Templ, Barbara; Koch, Elisabeth; Bolmgren, Kjell; Ungersböck, Markus; Paul, Anita; Scheifinger, Helfried; Rutishauser, This; Busto, Montserrat; Chmielewski, Frank-M.; Hájková, Lenka; Hodzić, Sabina; Kaspar, Frank; Pietragalla, Barbara; Romero-Fresneda, Ramiro; Tolvanen, Anne; Vučetič, Višnja; Zimmermann, Kirsten; Zust, Ana

    2018-02-01

    The Pan European Phenology (PEP) project is a European infrastructure to promote and facilitate phenological research, education, and environmental monitoring. The main objective is to maintain and develop a Pan European Phenological database (PEP725) with an open, unrestricted data access for science and education. PEP725 is the successor of the database developed through the COST action 725 "Establishing a European phenological data platform for climatological applications" working as a single access point for European-wide plant phenological data. So far, 32 European meteorological services and project partners from across Europe have joined and supplied data collected by volunteers from 1868 to the present for the PEP725 database. Most of the partners actively provide data on a regular basis. The database presently holds almost 12 million records, about 46 growing stages and 265 plant species (including cultivars), and can be accessed via http://www.pep725.eu/. Users of the PEP725 database have studied a diversity of topics ranging from climate change impact, plant physiological question, phenological modeling, and remote sensing of vegetation to ecosystem productivity.

  10. Pan European Phenological database (PEP725): a single point of access for European data

    Science.gov (United States)

    Templ, Barbara; Koch, Elisabeth; Bolmgren, Kjell; Ungersböck, Markus; Paul, Anita; Scheifinger, Helfried; Rutishauser, This; Busto, Montserrat; Chmielewski, Frank-M.; Hájková, Lenka; Hodzić, Sabina; Kaspar, Frank; Pietragalla, Barbara; Romero-Fresneda, Ramiro; Tolvanen, Anne; Vučetič, Višnja; Zimmermann, Kirsten; Zust, Ana

    2018-06-01

    The Pan European Phenology (PEP) project is a European infrastructure to promote and facilitate phenological research, education, and environmental monitoring. The main objective is to maintain and develop a Pan European Phenological database (PEP725) with an open, unrestricted data access for science and education. PEP725 is the successor of the database developed through the COST action 725 "Establishing a European phenological data platform for climatological applications" working as a single access point for European-wide plant phenological data. So far, 32 European meteorological services and project partners from across Europe have joined and supplied data collected by volunteers from 1868 to the present for the PEP725 database. Most of the partners actively provide data on a regular basis. The database presently holds almost 12 million records, about 46 growing stages and 265 plant species (including cultivars), and can be accessed via http://www.pep725.eu/ . Users of the PEP725 database have studied a diversity of topics ranging from climate change impact, plant physiological question, phenological modeling, and remote sensing of vegetation to ecosystem productivity.

  11. Calculation and analysis of the non-point source pollution in the upstream watershed of the Panjiakou Reservoir, People's Republic of China

    Science.gov (United States)

    Zhang, S.; Tang, L.

    2007-05-01

    Panjiakou Reservoir is an important drinking water resource in Haihe River Basin, Hebei Province, People's Republic of China. The upstream watershed area is about 35,000 square kilometers. Recently, the water pollution in the reservoir is becoming more serious owing to the non-point pollution as well as point source pollution on the upstream watershed. To effectively manage the reservoir and watershed and develop a plan to reduce pollutant loads, the loading of non-point and point pollution and their distribution on the upstream watershed must be understood fully. The SWAT model is used to simulate the production and transportation of the non-point source pollutants in the upstream watershed of the Panjiakou Reservoir. The loadings of non-point source pollutants are calculated for different hydrologic years and the spatial and temporal characteristics of non-point source pollution are studied. The stream network and topographic characteristics of the stream network and sub-basins are all derived from the DEM by ArcGIS software. The soil and land use data are reclassified and the soil physical properties database file is created for the model. The SWAT model was calibrated with observed data of several hydrologic monitoring stations in the study area. The results of the calibration show that the model performs fairly well. Then the calibrated model was used to calculate the loadings of non-point source pollutants for a wet year, a normal year and a dry year respectively. The time and space distribution of flow, sediment and non-point source pollution were analyzed depending on the simulated results. The comparison of different hydrologic years on calculation results is dramatic. The loading of non-point source pollution in the wet year is relatively larger but smaller in the dry year since the non-point source pollutants are mainly transported through the runoff. The pollution loading within a year is mainly produced in the flood season. Because SWAT is a

  12. Simulation of agricultural non-point source pollution in Xichuan by using SWAT model

    Science.gov (United States)

    Xing, Linan; Zuo, Jiane; Liu, Fenglin; Zhang, Xiaohui; Cao, Qiguang

    2018-02-01

    This paper evaluated the applicability of using SWAT to access agricultural non-point source pollution in Xichuan area. In order to build the model, DEM, soil sort and land use map, climate monitoring data were collected as basic database. The SWAT model was calibrated and validated for the SWAT was carried out using streamflow, suspended solids, total phosphorus and total nitrogen records from 2009 to 2011. Errors, coefficient of determination and Nash-Sutcliffe coefficient were considered to evaluate the applicability. The coefficient of determination were 0.96, 0.66, 0.55 and 0.66 for streamflow, SS, TN, and TP, respectively. Nash-Sutcliffe coefficient were 0.93, 0.5, 0.52 and 0.63, respectively. The results all meet the requirements. It suggested that the SWAT model can simulate the study area.

  13. ZeBase: an open-source relational database for zebrafish laboratories.

    Science.gov (United States)

    Hensley, Monica R; Hassenplug, Eric; McPhail, Rodney; Leung, Yuk Fai

    2012-03-01

    Abstract ZeBase is an open-source relational database for zebrafish inventory. It is designed for the recording of genetic, breeding, and survival information of fish lines maintained in a single- or multi-laboratory environment. Users can easily access ZeBase through standard web-browsers anywhere on a network. Convenient search and reporting functions are available to facilitate routine inventory work; such functions can also be automated by simple scripting. Optional barcode generation and scanning are also built-in for easy access to the information related to any fish. Further information of the database and an example implementation can be found at http://zebase.bio.purdue.edu.

  14. A method to analyze "source-sink" structure of non-point source pollution based on remote sensing technology.

    Science.gov (United States)

    Jiang, Mengzhen; Chen, Haiying; Chen, Qinghui

    2013-11-01

    With the purpose of providing scientific basis for environmental planning about non-point source pollution prevention and control, and improving the pollution regulating efficiency, this paper established the Grid Landscape Contrast Index based on Location-weighted Landscape Contrast Index according to the "source-sink" theory. The spatial distribution of non-point source pollution caused by Jiulongjiang Estuary could be worked out by utilizing high resolution remote sensing images. The results showed that, the area of "source" of nitrogen and phosphorus in Jiulongjiang Estuary was 534.42 km(2) in 2008, and the "sink" was 172.06 km(2). The "source" of non-point source pollution was distributed mainly over Xiamen island, most of Haicang, east of Jiaomei and river bank of Gangwei and Shima; and the "sink" was distributed over southwest of Xiamen island and west of Shima. Generally speaking, the intensity of "source" gets weaker along with the distance from the seas boundary increase, while "sink" gets stronger. Copyright © 2013 Elsevier Ltd. All rights reserved.

  15. Trans-Z-source and Γ-Z-source neutral-point-clamped inverters

    DEFF Research Database (Denmark)

    Wei, Mo; Loh, Poh Chiang; Blaabjerg, Frede

    2015-01-01

    Z-source neutral-point-clamped (NPC) inverters are earlier proposed for obtaining voltage buck-boost and three-level switching simultaneously. Their performances are, however, constrained by a trade-off between their input-to-output gain and modulation ratio. This trade-off can lead to high...

  16. Ibmdbpy-spatial : An Open-source implementation of in-database geospatial analytics in Python

    Science.gov (United States)

    Roy, Avipsa; Fouché, Edouard; Rodriguez Morales, Rafael; Moehler, Gregor

    2017-04-01

    As the amount of spatial data acquired from several geodetic sources has grown over the years and as data infrastructure has become more powerful, the need for adoption of in-database analytic technology within geosciences has grown rapidly. In-database analytics on spatial data stored in a traditional enterprise data warehouse enables much faster retrieval and analysis for making better predictions about risks and opportunities, identifying trends and spot anomalies. Although there are a number of open-source spatial analysis libraries like geopandas and shapely available today, most of them have been restricted to manipulation and analysis of geometric objects with a dependency on GEOS and similar libraries. We present an open-source software package, written in Python, to fill the gap between spatial analysis and in-database analytics. Ibmdbpy-spatial provides a geospatial extension to the ibmdbpy package, implemented in 2015. It provides an interface for spatial data manipulation and access to in-database algorithms in IBM dashDB, a data warehouse platform with a spatial extender that runs as a service on IBM's cloud platform called Bluemix. Working in-database reduces the network overload, as the complete data need not be replicated into the user's local system altogether and only a subset of the entire dataset can be fetched into memory in a single instance. Ibmdbpy-spatial accelerates Python analytics by seamlessly pushing operations written in Python into the underlying database for execution using the dashDB spatial extender, thereby benefiting from in-database performance-enhancing features, such as columnar storage and parallel processing. The package is currently supported on Python versions from 2.7 up to 3.4. The basic architecture of the package consists of three main components - 1) a connection to the dashDB represented by the instance IdaDataBase, which uses a middleware API namely - pypyodbc or jaydebeapi to establish the database connection via

  17. MyMolDB: a micromolecular database solution with open source and free components.

    Science.gov (United States)

    Xia, Bing; Tai, Zheng-Fu; Gu, Yu-Cheng; Li, Bang-Jing; Ding, Li-Sheng; Zhou, Yan

    2011-10-01

    To manage chemical structures in small laboratories is one of the important daily tasks. Few solutions are available on the internet, and most of them are closed source applications. The open-source applications typically have limited capability and basic cheminformatics functionalities. In this article, we describe an open-source solution to manage chemicals in research groups based on open source and free components. It has a user-friendly interface with the functions of chemical handling and intensive searching. MyMolDB is a micromolecular database solution that supports exact, substructure, similarity, and combined searching. This solution is mainly implemented using scripting language Python with a web-based interface for compound management and searching. Almost all the searches are in essence done with pure SQL on the database by using the high performance of the database engine. Thus, impressive searching speed has been archived in large data sets for no external Central Processing Unit (CPU) consuming languages were involved in the key procedure of the searching. MyMolDB is an open-source software and can be modified and/or redistributed under GNU General Public License version 3 published by the Free Software Foundation (Free Software Foundation Inc. The GNU General Public License, Version 3, 2007. Available at: http://www.gnu.org/licenses/gpl.html). The software itself can be found at http://code.google.com/p/mymoldb/. Copyright © 2011 Wiley Periodicals, Inc.

  18. Analysis of commercial and public bioactivity databases.

    Science.gov (United States)

    Tiikkainen, Pekka; Franke, Lutz

    2012-02-27

    Activity data for small molecules are invaluable in chemoinformatics. Various bioactivity databases exist containing detailed information of target proteins and quantitative binding data for small molecules extracted from journals and patents. In the current work, we have merged several public and commercial bioactivity databases into one bioactivity metabase. The molecular presentation, target information, and activity data of the vendor databases were standardized. The main motivation of the work was to create a single relational database which allows fast and simple data retrieval by in-house scientists. Second, we wanted to know the amount of overlap between databases by commercial and public vendors to see whether the former contain data complementing the latter. Third, we quantified the degree of inconsistency between data sources by comparing data points derived from the same scientific article cited by more than one vendor. We found that each data source contains unique data which is due to different scientific articles cited by the vendors. When comparing data derived from the same article we found that inconsistencies between the vendors are common. In conclusion, using databases of different vendors is still useful since the data overlap is not complete. It should be noted that this can be partially explained by the inconsistencies and errors in the source data.

  19. BEAMLINE-CONTROLLED STEERING OF SOURCE-POINT ANGLE AT THE ADVANCED PHOTON SOURCE

    Energy Technology Data Exchange (ETDEWEB)

    Emery, L.; Fystro, G.; Shang, H.; Smith, M.

    2017-06-25

    An EPICS-based steering software system has been implemented for beamline personnel to directly steer the angle of the synchrotron radiation sources at the Advanced Photon Source. A script running on a workstation monitors "start steering" beamline EPICS records, and effects a steering given by the value of the "angle request" EPICS record. The new system makes the steering process much faster than before, although the older steering protocols can still be used. The robustness features of the original steering remain. Feedback messages are provided to the beamlines and the accelerator operators. Underpinning this new steering protocol is the recent refinement of the global orbit feedback process whereby feedforward of dipole corrector set points and orbit set points are used to create a local steering bump in a rapid and seamless way.

  20. Isotropic irradiation of detectors from point sources

    DEFF Research Database (Denmark)

    Aage, Helle Karina

    1997-01-01

    NaI(Tl) scintillator detectors have been exposed to gamma rays from 8 different point sources from different directions. Background and backscatter of gamma-rays from the surroundings have been subtracted in order to produce clean spectra. By adding spectra obtained from exposures from different ...

  1. IAEA Illicit Trafficking Database (ITDB)

    International Nuclear Information System (INIS)

    2010-01-01

    The IAEA Illicit Trafficking Database (ITDB) was established in 1995 as a unique network of points of contact connecting 100 states and several international organizations. Information collected from official sources supplemented by open-source reports. The 1994 - GC 38, resolution intensifies the activities through which the Agency is currently supporting Member States in this field. Member states were notified of completed database in 1995 and invited to participate. The purpose of the I TDB is to facilitate exchange of authoritative information among States on incidents of illicit trafficking and other related unauthorized activities involving nuclear and other radioactive materials; to collect, maintain and analyse information on such incidents with a view to identifying common threats, trends, and patterns; use this information for internal planning and prioritisation and provide this information to member states and to provide a reliable source of basic information on such incidents to the media, when appropriate

  2. Thermal Analysis of a Cracked Half-plane under Moving Point Heat Source

    Directory of Open Access Journals (Sweden)

    He Kuanfang

    2017-09-01

    Full Text Available The heat conduction in half-plane with an insulated crack subjected to moving point heat source is investigated. The analytical solution and the numerical means are combined to analyze the transient temperature distribution of a cracked half-plane under moving point heat source. The transient temperature distribution of the half plane structure under moving point heat source is obtained by the moving coordinate method firstly, then the heat conduction equation with thermal boundary of an insulated crack face is changed to singular integral equation by applying Fourier transforms and solved by the numerical method. The numerical examples of the temperature distribution on the cracked half-plane structure under moving point heat source are presented and discussed in detail.

  3. Search for high energy cosmic neutrino point sources with ANTARES

    International Nuclear Information System (INIS)

    Halladjian, G.

    2010-01-01

    The aim of this thesis is the search for high energy cosmic neutrinos emitted by point sources with the ANTARES neutrino telescope. The detection of high energy cosmic neutrinos can bring answers to important questions such as the origin of cosmic rays and the γ-rays emission processes. In the first part of the thesis, the neutrino flux emitted by galactic and extragalactic sources and the number of events which can be detected by ANTARES are estimated. This study uses the measured γ-ray spectra of known sources taking into account the γ-ray absorption by the extragalactic background light. In the second part of the thesis, the absolute pointing of the ANTARES telescope is evaluated. Being located at a depth of 2475 m in sea water, the orientation of the detector is determined by an acoustic positioning system which relies on low and high frequency acoustic waves measurements between the sea surface and the bottom. The third part of the thesis is a search for neutrino point sources in the ANTARES data. The search algorithm is based on a likelihood ratio maximization method. It is used in two search strategies; 'the candidate sources list strategy' and 'the all sky search strategy'. Analysing 2007+2008 data, no discovery is made and the world's best upper limits on neutrino fluxes from various sources in the Southern sky are established. (author)

  4. Nomogram for Determining Shield Thickness for Point and Line Sources of Gamma Rays

    Energy Technology Data Exchange (ETDEWEB)

    Joenemalm, C; Malen, K

    1966-10-15

    A set of nomograms is given for the determination of the required shield thickness against gamma radiation. The sources handled are point and infinite line sources with shields of Pb, Fe, magnetite concrete (p = 3.6), ordinary concrete (p = 2.3) or water. The gamma energy range covered is 0.5 - 10 MeV. The nomograms are directly applicable for source and dose points on the surfaces of the shield. They can easily be extended to source and dose points in other positions by applying a geometrical correction. Also included are data for calculation of the source strength for the most common materials and for fission product sources.

  5. Nomogram for Determining Shield Thickness for Point and Line Sources of Gamma Rays

    International Nuclear Information System (INIS)

    Joenemalm, C.; Malen, K

    1966-10-01

    A set of nomograms is given for the determination of the required shield thickness against gamma radiation. The sources handled are point and infinite line sources with shields of Pb, Fe, magnetite concrete (p = 3.6), ordinary concrete (p = 2.3) or water. The gamma energy range covered is 0.5 - 10 MeV. The nomograms are directly applicable for source and dose points on the surfaces of the shield. They can easily be extended to source and dose points in other positions by applying a geometrical correction. Also included are data for calculation of the source strength for the most common materials and for fission product sources

  6. Point source search techniques in ultra high energy gamma ray astronomy

    International Nuclear Information System (INIS)

    Alexandreas, D.E.; Biller, S.; Dion, G.M.; Lu, X.Q.; Yodh, G.B.; Berley, D.; Goodman, J.A.; Haines, T.J.; Hoffman, C.M.; Horch, E.; Sinnis, C.; Zhang, W.

    1993-01-01

    Searches for point astrophysical sources of ultra high energy (UHE) gamma rays are plagued by large numbers of background events from isotropic cosmic rays. Some of the methods that have been used to estimate the expected number of background events coming from the direction of a possible source are found to contain biases. Search techniques that avoid this problem are described. There is also a discussion of how to optimize the sensitivity of a search to emission from a point source. (orig.)

  7. Open-Source Automated Mapping Four-Point Probe

    Directory of Open Access Journals (Sweden)

    Handy Chandra

    2017-01-01

    Full Text Available Scientists have begun using self-replicating rapid prototyper (RepRap 3-D printers to manufacture open source digital designs of scientific equipment. This approach is refined here to develop a novel instrument capable of performing automated large-area four-point probe measurements. The designs for conversion of a RepRap 3-D printer to a 2-D open source four-point probe (OS4PP measurement device are detailed for the mechanical and electrical systems. Free and open source software and firmware are developed to operate the tool. The OS4PP was validated against a wide range of discrete resistors and indium tin oxide (ITO samples of different thicknesses both pre- and post-annealing. The OS4PP was then compared to two commercial proprietary systems. Results of resistors from 10 to 1 MΩ show errors of less than 1% for the OS4PP. The 3-D mapping of sheet resistance of ITO samples successfully demonstrated the automated capability to measure non-uniformities in large-area samples. The results indicate that all measured values are within the same order of magnitude when compared to two proprietary measurement systems. In conclusion, the OS4PP system, which costs less than 70% of manual proprietary systems, is comparable electrically while offering automated 100 micron positional accuracy for measuring sheet resistance over larger areas.

  8. Open-Source Automated Mapping Four-Point Probe.

    Science.gov (United States)

    Chandra, Handy; Allen, Spencer W; Oberloier, Shane W; Bihari, Nupur; Gwamuri, Jephias; Pearce, Joshua M

    2017-01-26

    Scientists have begun using self-replicating rapid prototyper (RepRap) 3-D printers to manufacture open source digital designs of scientific equipment. This approach is refined here to develop a novel instrument capable of performing automated large-area four-point probe measurements. The designs for conversion of a RepRap 3-D printer to a 2-D open source four-point probe (OS4PP) measurement device are detailed for the mechanical and electrical systems. Free and open source software and firmware are developed to operate the tool. The OS4PP was validated against a wide range of discrete resistors and indium tin oxide (ITO) samples of different thicknesses both pre- and post-annealing. The OS4PP was then compared to two commercial proprietary systems. Results of resistors from 10 to 1 MΩ show errors of less than 1% for the OS4PP. The 3-D mapping of sheet resistance of ITO samples successfully demonstrated the automated capability to measure non-uniformities in large-area samples. The results indicate that all measured values are within the same order of magnitude when compared to two proprietary measurement systems. In conclusion, the OS4PP system, which costs less than 70% of manual proprietary systems, is comparable electrically while offering automated 100 micron positional accuracy for measuring sheet resistance over larger areas.

  9. Scattering and absorption of particles emitted by a point source in a cluster of point scatterers

    International Nuclear Information System (INIS)

    Liljequist, D.

    2012-01-01

    A theory for the scattering and absorption of particles isotropically emitted by a point source in a cluster of point scatterers is described and related to the theory for the scattering of an incident particle beam. The quantum mechanical probability of escape from the cluster in different directions is calculated, as well as the spatial distribution of absorption events within the cluster. A source strength renormalization procedure is required. The average quantum scattering in clusters with randomly shifting scatterer positions is compared to trajectory simulation with the aim of studying the validity of the trajectory method. Differences between the results of the quantum and trajectory methods are found primarily for wavelengths larger than the average distance between nearest neighbour scatterers. The average quantum results include, for example, a local minimum in the number of absorption events at the location of the point source and interference patterns in the angle-dependent escape probability as well as in the distribution of absorption events. The relative error of the trajectory method is in general, though not generally, of similar magnitude as that obtained for beam scattering.

  10. OH masers associated with IRAS point sources

    NARCIS (Netherlands)

    Masheder, MRW; Cohen, RJ; Martin-Hernandez, NL; Migenes,; Reid, MJ

    2002-01-01

    We report a search for masers from the Lambda-doublet of the ground-state of OH at 18cm, carried out with the Jodrell Bank Lovell Telescope and with the 25m Dwingeloo telescope. All objects north of delta = -20degrees which appear in the IRAS Point Source Catalog with fluxes > 1000 Jy at 60mum and

  11. Resolution of point sources of light as analyzed by quantum detection theory.

    Science.gov (United States)

    Helstrom, C. W.

    1973-01-01

    The resolvability of point sources of incoherent thermal light is analyzed by quantum detection theory in terms of two hypothesis-testing problems. In the first, the observer must decide whether there are two sources of equal radiant power at given locations, or whether there is only one source of twice the power located midway between them. In the second problem, either one, but not both, of two point sources is radiating, and the observer must decide which it is. The decisions are based on optimum processing of the electromagnetic field at the aperture of an optical instrument. In both problems the density operators of the field under the two hypotheses do not commute. The error probabilities, determined as functions of the separation of the points and the mean number of received photons, characterize the ultimate resolvability of the sources.

  12. Mapping correlation of a simulated dark matter source and a point source in the gamma-ray sky - Oral Presentation

    Energy Technology Data Exchange (ETDEWEB)

    Gibson, Alexander [SLAC National Accelerator Lab., Menlo Park, CA (United States)

    2015-08-23

    In my research, I analyzed how two gamma-ray source models interact with one another when optimizing to fit data. This is important because it becomes hard to distinguish between the two point sources when they are close together or looking at low energy photons. The reason for the first is obvious, the reason why they become harder to distinguish at lower photon energies is the resolving power of the Fermi Gamma-Ray Space Telescope gets worse at lower energies. When the two point sources are highly correlated (hard to distinguish between), we need to change our method of statistical analysis. What I did was show that highly correlated sources have larger uncertainties associated with them, caused by an optimizer not knowing which point source’s parameters to optimize. I also mapped out where their is high correlation for 2 different theoretical mass dark matter point sources so that people analyzing them in the future knew where they had to use more sophisticated statistical analysis.

  13. Localization of Point Sources for Poisson Equation using State Observers

    KAUST Repository

    Majeed, Muhammad Usman

    2016-08-09

    A method based On iterative observer design is presented to solve point source localization problem for Poisson equation with riven boundary data. The procedure involves solution of multiple boundary estimation sub problems using the available Dirichlet and Neumann data from different parts of the boundary. A weighted sum of these solution profiles of sub-problems localizes point sources inside the domain. Method to compute these weights is also provided. Numerical results are presented using finite differences in a rectangular domain. (C) 2016, IFAC (International Federation of Automatic Control) Hosting by Elsevier Ltd. All rights reserved.

  14. Localization of Point Sources for Poisson Equation using State Observers

    KAUST Repository

    Majeed, Muhammad Usman; Laleg-Kirati, Taous-Meriem

    2016-01-01

    A method based On iterative observer design is presented to solve point source localization problem for Poisson equation with riven boundary data. The procedure involves solution of multiple boundary estimation sub problems using the available Dirichlet and Neumann data from different parts of the boundary. A weighted sum of these solution profiles of sub-problems localizes point sources inside the domain. Method to compute these weights is also provided. Numerical results are presented using finite differences in a rectangular domain. (C) 2016, IFAC (International Federation of Automatic Control) Hosting by Elsevier Ltd. All rights reserved.

  15. Point sources and multipoles in inverse scattering theory

    CERN Document Server

    Potthast, Roland

    2001-01-01

    Over the last twenty years, the growing availability of computing power has had an enormous impact on the classical fields of direct and inverse scattering. The study of inverse scattering, in particular, has developed rapidly with the ability to perform computational simulations of scattering processes and led to remarkable advances in a range of applications, from medical imaging and radar to remote sensing and seismic exploration. Point Sources and Multipoles in Inverse Scattering Theory provides a survey of recent developments in inverse acoustic and electromagnetic scattering theory. Focusing on methods developed over the last six years by Colton, Kirsch, and the author, this treatment uses point sources combined with several far-reaching techniques to obtain qualitative reconstruction methods. The author addresses questions of uniqueness, stability, and reconstructions for both two-and three-dimensional problems.With interest in extracting information about an object through scattered waves at an all-ti...

  16. The peak efficiency calibration of volume source using 152Eu point source in computer

    International Nuclear Information System (INIS)

    Shen Tingyun; Qian Jianfu; Nan Qinliang; Zhou Yanguo

    1997-01-01

    The author describes the method of the peak efficiency calibration of volume source by means of 152 Eu point source for HPGe γ spectrometer. The peak efficiency can be computed by Monte Carlo simulation, after inputting parameter of detector. The computation results are in agreement with the experimental results with an error of +-3.8%, with an exception one is about +-7.4%

  17. Kajian Unified Theory of Acceptance and Use of Technology Dalam Penggunaan Open Source Software Database Management System

    Directory of Open Access Journals (Sweden)

    Michael Sonny

    2016-06-01

    Full Text Available Perkembangan perangkat lunak computer dewasa ini terjadi sedemikian pesatnya, perkembangan tidak hanya terjadi pada perangkat lunak yang memiliki lisensi tertentu, perangkat open source pun demikian. Perkembangan itu tentu saja sangat menggembirakan bagi pengguna computer khususnya di kalangan pendidikan maupun di kalangan mahasiswa, karena pengguna mempunyai beberapa pilihan untuk menggunakan aplikasi. Perangkat lunak open source juga menawarkan produk yang umumnya gratis, diberikan kode programnya, kebebasan untuk modifikasi dan mengembangkan. Meneliti aplikasi berbasis open source tentu saja sangat beragam seperti aplikasi untuk pemrograman (PHP, Gambas, Database Management System (MySql, SQLite, browsing (Mozilla, Firefox, Opera. Pada penelitian ini di kaji penerimaan aplikasi DBMS (Database Management System seperti MySql dan SQLite dengan menggunakan sebuah model yang dikembangkan oleh Venkantes(2003 yaitu UTAUT (Unified Theory of Acceptance and Use of Technology. Faktor – faktor tertentu juga mempengaruhi dalam melakukan kegiatan pembelajaran aplikasi open source ini, salah satu faktor atau yang disebut dengan moderating yang bisa mempengaruhi efektifitas dan efisiensi. Dengan demikian akan mendapatkan hasil yang bisa membuat kelancaran dalam pembelajaran aplikasi berbasis open source ini.   Kata kunci— open source, Database Management System (DBMS, Modereting

  18. Lessons Learned from resolving massive IPS database change for SPADES+

    International Nuclear Information System (INIS)

    Kim, Jin-Soo

    2016-01-01

    Safety Parameter Display and Evaluation System+ (SPADES+) was implemented to meet the requirements for Safety Parameter Display System (SPDS) which are related to TMI Action Plan requirements. SPADES+ monitors continuously the critical safety function during normal, abnormal, and emergency operation mode and generates the alarm output to the alarm server when the tolerance related to safety functions are not satisfied. The alarm algorithm for critical safety function is performed in the NSSS Application Software (NAPS) server of the Information Process System (IPS) and the calculation result will be displayed on the flat panel display (FPD) of the IPS. SPADES+ provides the critical variable to the control room operators to aid them in rapidly and reliable determining the safety status of the plant. Many database point ID names (518 points) were changed. POINT_ID is used in the programming source code, the related documents such as SDS and SRS, and Graphic database. To reduce human errors, computer program and office program’s Macro are used. Though the automatic methods are used for changing POINT_IDs, it takes lots of time to resolve for editing the change list except for making computerized solutions. In IPS, there are many more programs than SPADES+ and over 30,000 POINT_IDs are in IPS database. Changing POINT_IDs could be a burden to software engineers. In case of Ovation system database, there is the Alias field to prevent this kind of problem. The Alias is a kind of secondary key in database

  19. Optical identifications of IRAS point sources: the Fornax, Hydra I and Coma clusters

    International Nuclear Information System (INIS)

    Wang, G.; Leggett, S.K.; Savage, A.

    1991-01-01

    We present optical identifications for 66 IRAS point sources in the region of the Fornax cluster of galaxies, 106 IRAS point sources in the region of the Hydra I cluster of galaxies (Abell 1060) and 59 IRAS point sources in the region of the Coma cluster of galaxies (Abell 1656). Eight other sources in Hydra I do not have optical counterparts and are very probably due to infrared cirrus. Twenty-three (35 per cent) of the Fornax sources are associated with stars and 43 (65 per cent) with galaxies; 48 (42 per cent) of the Hydra I sources are associated with stars and 58 (51 per cent) with galaxies; 18 (31 per cent) of the Coma sources are associated with stars and 41 (69 per cent) with galaxies. The stellar and infrared cirrus surface density is consistent with the galactic latitude of each field. (author)

  20. LOWERING ICECUBE'S ENERGY THRESHOLD FOR POINT SOURCE SEARCHES IN THE SOUTHERN SKY

    Energy Technology Data Exchange (ETDEWEB)

    Aartsen, M. G. [Department of Physics, University of Adelaide, Adelaide, 5005 (Australia); Abraham, K. [Physik-department, Technische Universität München, D-85748 Garching (Germany); Ackermann, M. [DESY, D-15735 Zeuthen (Germany); Adams, J. [Department of Physics and Astronomy, University of Canterbury, Private Bag 4800, Christchurch (New Zealand); Aguilar, J. A.; Ansseau, I. [Université Libre de Bruxelles, Science Faculty CP230, B-1050 Brussels (Belgium); Ahlers, M. [Department of Physics and Wisconsin IceCube Particle Astrophysics Center, University of Wisconsin, Madison, WI 53706 (United States); Ahrens, M. [Oskar Klein Centre and Department of Physics, Stockholm University, SE-10691 Stockholm (Sweden); Altmann, D.; Anton, G. [Erlangen Centre for Astroparticle Physics, Friedrich-Alexander-Universität Erlangen-Nürnberg, D-91058 Erlangen (Germany); Andeen, K. [Department of Physics, Marquette University, Milwaukee, WI, 53201 (United States); Anderson, T.; Arlen, T. C. [Department of Physics, Pennsylvania State University, University Park, PA 16802 (United States); Archinger, M.; Baum, V. [Institute of Physics, University of Mainz, Staudinger Weg 7, D-55099 Mainz (Germany); Arguelles, C. [Department of Physics, Massachusetts Institute of Technology, Cambridge, MA 02139 (United States); Auffenberg, J. [III. Physikalisches Institut, RWTH Aachen University, D-52056 Aachen (Germany); Bai, X. [Physics Department, South Dakota School of Mines and Technology, Rapid City, SD 57701 (United States); Barwick, S. W. [Department of Physics and Astronomy, University of California, Irvine, CA 92697 (United States); Bay, R., E-mail: jacob.feintzeig@gmail.com, E-mail: naoko@icecube.wisc.edu [Department of Physics, University of California, Berkeley, CA 94720 (United States); Collaboration: IceCube Collaboration; and others

    2016-06-20

    Observation of a point source of astrophysical neutrinos would be a “smoking gun” signature of a cosmic-ray accelerator. While IceCube has recently discovered a diffuse flux of astrophysical neutrinos, no localized point source has been observed. Previous IceCube searches for point sources in the southern sky were restricted by either an energy threshold above a few hundred TeV or poor neutrino angular resolution. Here we present a search for southern sky point sources with greatly improved sensitivities to neutrinos with energies below 100 TeV. By selecting charged-current ν{sub μ} interacting inside the detector, we reduce the atmospheric background while retaining efficiency for astrophysical neutrino-induced events reconstructed with sub-degree angular resolution. The new event sample covers three years of detector data and leads to a factor of 10 improvement in sensitivity to point sources emitting below 100 TeV in the southern sky. No statistically significant evidence of point sources was found, and upper limits are set on neutrino emission from individual sources. A posteriori analysis of the highest-energy (∼100 TeV) starting event in the sample found that this event alone represents a 2.8 σ deviation from the hypothesis that the data consists only of atmospheric background.

  1. jSPyDB, an open source database-independent tool for data management

    CERN Document Server

    Pierro, Giuseppe Antonio

    2010-01-01

    Nowadays, the number of commercial tools available for accessing Databases, built on Java or .Net, is increasing. However, many of these applications have several drawbacks: usually they are not open-source, they provide interfaces only with a specific kind of database, they are platform-dependent and very CPU and memory consuming. jSPyDB is a free web based tool written using Python and Javascript. It relies on jQuery and python libraries, and is intended to provide a simple handler to different Database technologies inside a local web browser. Such a tool, exploiting fast access libraries such as SQLAlchemy, is easy to install, and to configure. The design of this tool envisages three layers. The front-end client side in the local web browser communicates with a backend server. Only the server is able to connect to the different databases for the purposes of performing data definition and manipulation. The server makes the data available to the client, so that the user can display and handle them safely. ...

  2. Mars Science Laboratory Frame Manager for Centralized Frame Tree Database and Target Pointing

    Science.gov (United States)

    Kim, Won S.; Leger, Chris; Peters, Stephen; Carsten, Joseph; Diaz-Calderon, Antonio

    2013-01-01

    The FM (Frame Manager) flight software module is responsible for maintaining the frame tree database containing coordinate transforms between frames. The frame tree is a proper tree structure of directed links, consisting of surface and rover subtrees. Actual frame transforms are updated by their owner. FM updates site and saved frames for the surface tree. As the rover drives to a new area, a new site frame with an incremented site index can be created. Several clients including ARM and RSM (Remote Sensing Mast) update their related rover frames that they own. Through the onboard centralized FM frame tree database, client modules can query transforms between any two frames. Important applications include target image pointing for RSM-mounted cameras and frame-referenced arm moves. The use of frame tree eliminates cumbersome, error-prone calculations of coordinate entries for commands and thus simplifies flight operations significantly.

  3. Performance of popular open source databases for HEP related computing problems

    International Nuclear Information System (INIS)

    Kovalskyi, D; Sfiligoi, I; Wuerthwein, F; Yagil, A

    2014-01-01

    Databases are used in many software components of HEP computing, from monitoring and job scheduling to data storage and processing. It is not always clear at the beginning of a project if a problem can be handled by a single server, or if one needs to plan for a multi-server solution. Before a scalable solution is adopted, it helps to know how well it performs in a single server case to avoid situations when a multi-server solution is adopted mostly due to sub-optimal performance per node. This paper presents comparison benchmarks of popular open source database management systems. As a test application we use a user job monitoring system based on the Glidein workflow management system used in the CMS Collaboration.

  4. Tackling non-point source water pollution in British Columbia: An action plan

    Energy Technology Data Exchange (ETDEWEB)

    1998-01-01

    Efforts to protect British Columbia water quality by regulating point discharges from municipal and industrial sources have generally been successful, and it is recognized that the major remaining cause of water pollution in the province is from non-point sources. These sources are largely unregulated and associated with urbanization, agriculture, and other forms of land development. The first part of this report reviews the provincial commitment to clean water, the effects of non-point-source (NPS) pollution, and the management of NPS in the province. Part 2 describes the main causes of NPS in British Columbia: Land development, agriculture, stormwater runoff, on-site sewage systems, forestry and range activities, atmospheric deposition, and boating/marine activities. Finally, it presents key components of the province's NPS action plan: Education and training, prevention at site, land use planning and co-ordination, assessment and reporting, economic incentives, legislation and regulation, and implementation.

  5. Epidemiology, public health, and health surveillance around point sources of pollution

    International Nuclear Information System (INIS)

    Stebbings, J.H. Jr.

    1981-01-01

    In industrial society a large number of point sources of pollution exist, such as chemical plants, smelters, and nuclear power plants. Public concern has forced the practising epidemiologist to undertake health surveillance of the usually small populations living around point sources. Although not justifiable as research, such epidemiologic surveillance activities are becoming a routine part of public health practice, and this trend will continue. This introduction reviews concepts of epidemiologic surveillance, and institutional problems relating to the quality of such applied research

  6. The resolution of point sources of light as analyzed by quantum detection theory

    Science.gov (United States)

    Helstrom, C. W.

    1972-01-01

    The resolvability of point sources of incoherent light is analyzed by quantum detection theory in terms of two hypothesis-testing problems. In the first, the observer must decide whether there are two sources of equal radiant power at given locations, or whether there is only one source of twice the power located midway between them. In the second problem, either one, but not both, of two point sources is radiating, and the observer must decide which it is. The decisions are based on optimum processing of the electromagnetic field at the aperture of an optical instrument. In both problems the density operators of the field under the two hypotheses do not commute. The error probabilities, determined as functions of the separation of the points and the mean number of received photons, characterize the ultimate resolvability of the sources.

  7. Interferometry with flexible point source array for measuring complex freeform surface and its design algorithm

    Science.gov (United States)

    Li, Jia; Shen, Hua; Zhu, Rihong; Gao, Jinming; Sun, Yue; Wang, Jinsong; Li, Bo

    2018-06-01

    The precision of the measurements of aspheric and freeform surfaces remains the primary factor restrict their manufacture and application. One effective means of measuring such surfaces involves using reference or probe beams with angle modulation, such as tilted-wave-interferometer (TWI). It is necessary to improve the measurement efficiency by obtaining the optimum point source array for different pieces before TWI measurements. For purpose of forming a point source array based on the gradients of different surfaces under test, we established a mathematical model describing the relationship between the point source array and the test surface. However, the optimal point sources are irregularly distributed. In order to achieve a flexible point source array according to the gradient of test surface, a novel interference setup using fiber array is proposed in which every point source can be independently controlled on and off. Simulations and the actual measurement examples of two different surfaces are given in this paper to verify the mathematical model. Finally, we performed an experiment of testing an off-axis ellipsoidal surface that proved the validity of the proposed interference system.

  8. Lessons Learned from resolving massive IPS database change for SPADES+

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jin-Soo [KEPCO Engineering and Construction Co., Deajeon (Korea, Republic of)

    2016-10-15

    Safety Parameter Display and Evaluation System+ (SPADES+) was implemented to meet the requirements for Safety Parameter Display System (SPDS) which are related to TMI Action Plan requirements. SPADES+ monitors continuously the critical safety function during normal, abnormal, and emergency operation mode and generates the alarm output to the alarm server when the tolerance related to safety functions are not satisfied. The alarm algorithm for critical safety function is performed in the NSSS Application Software (NAPS) server of the Information Process System (IPS) and the calculation result will be displayed on the flat panel display (FPD) of the IPS. SPADES+ provides the critical variable to the control room operators to aid them in rapidly and reliable determining the safety status of the plant. Many database point ID names (518 points) were changed. POINT{sub I}D is used in the programming source code, the related documents such as SDS and SRS, and Graphic database. To reduce human errors, computer program and office program’s Macro are used. Though the automatic methods are used for changing POINT{sub I}Ds, it takes lots of time to resolve for editing the change list except for making computerized solutions. In IPS, there are many more programs than SPADES+ and over 30,000 POINT{sub I}Ds are in IPS database. Changing POINT{sub I}Ds could be a burden to software engineers. In case of Ovation system database, there is the Alias field to prevent this kind of problem. The Alias is a kind of secondary key in database.

  9. Model Predictive Control of Z-source Neutral Point Clamped Inverter

    DEFF Research Database (Denmark)

    Mo, Wei; Loh, Poh Chiang; Blaabjerg, Frede

    2011-01-01

    This paper presents Model Predictive Control (MPC) of Z-source Neutral Point Clamped (NPC) inverter. For illustration, current control of Z-source NPC grid-connected inverter is analyzed and simulated. With MPC’s advantage of easily including system constraints, load current, impedance network...... response are obtained at the same time with a formulated Z-source NPC inverter network model. Operation steady state and transient state simulation results of MPC are going to be presented, which shows good reference tracking ability of this method. It provides new control method for Z-source NPC inverter...

  10. Opinions on Drug Interaction Sources in Anticancer Treatments and Parameters for an Oncology-Specific Database by Pharmacy Practitioners in Asia

    Directory of Open Access Journals (Sweden)

    2010-01-01

    Full Text Available Cancer patients undergoing chemotherapy are particularly susceptible to drug-drug interactions (DDIs. Practitioners should keep themselves updated with the most current DDI information, particularly involving new anticancer drugs (ACDs. Databases can be useful to obtain up-to-date DDI information in a timely and efficient manner. Our objective was to investigate the DDI information sources of pharmacy practitioners in Asia and their views on the usefulness of an oncology-specific database for ACD interactions. A qualitative, cross-sectional survey was done to collect information on the respondents' practice characteristics, sources of DDI information and parameters useful in an ACD interaction database. Response rate was 49%. Electronic databases (70%, drug interaction textbooks (69% and drug compendia (64% were most commonly used. Majority (93% indicated that a database catering towards ACD interactions was useful. Essential parameters that should be included in the database were the mechanism and severity of the detected interaction, and the presence of a management plan (98% each. This study has improved our understanding on the usefulness of various DDI information sources for ACD interactions among pharmacy practitioners in Asia. An oncology-specific DDI database targeting ACD interactions is definitely attractive for clinical practice.

  11. Pulsewidth-modulated 2-source neutral-point-clamped inverter

    DEFF Research Database (Denmark)

    Blaabjerg, Frede; Loh, Poh Chang; Gao, Feng

    2007-01-01

    This paper presents the careful integration of a newly proposed Z-source topological concept to the basic neutral-point-clamped (NPC) inverter topology for designing a three-level inverter with both voltage-buck and voltage-boost capabilities. The designed Z-source NPC inverter uses two unique X......-shaped inductance-capacitance (LC) impedance networks that are connected between two isolated dc input power sources and its inverter circuitry for boosting its AC output voltage. Through the design of an appropriate pulsewidth-modulation (PWM) algorithm, the two impedance networks can be short......-circuited sequentially (without shooting through the inverter full DC link) for implementing the ldquonearest-three-vectorrdquo modulation principle with minimized harmonic distortion and device commutations per half carrier cycle while performing voltage boosting. With only a slight modification to the inverter PWM...

  12. Clinical Validation of Point-Source Corneal Topography in Keratoplasty

    NARCIS (Netherlands)

    Vrijling, A C L; Braaf, B.; Snellenburg, J.J.; de Lange, F.; Zaal, M.J.W.; van der Heijde, G.L.; Sicam, V.A.D.P.

    2011-01-01

    Purpose. To validate the clinical performance of point-source corneal topography (PCT) in postpenetrating keratoplasty (PKP) eyes and to compare it with conventional Placido-based topography. Methods. Corneal elevation maps of the anterior corneal surface were obtained from 20 post-PKP corneas using

  13. PSD Applicability Determination for Multiple Owner/Operator Point Sources Within a Single Facility

    Science.gov (United States)

    This document may be of assistance in applying the New Source Review (NSR) air permitting regulations including the Prevention of Significant Deterioration (PSD) requirements. This document is part of the NSR Policy and Guidance Database. Some documents in the database are a scanned or retyped version of a paper photocopy of the original. Although we have taken considerable effort to quality assure the documents, some may contain typographical errors. Contact the office that issued the document if you need a copy of the original.

  14. Characteristics of a multi-keV monochromatic point x-ray source

    Indian Academy of Sciences (India)

    Temporal, spatial and spectral characteristics of a multi-keV monochromatic point x-ray source based on vacuum diode with laser-produced plasma as cathode are presented. Electrons from a laser-produced aluminium plasma were accelerated towards a conical point tip titanium anode to generate K-shell x-ray radiation.

  15. Fast in-database cross-matching of high-cadence, high-density source lists with an up-to-date sky model

    Science.gov (United States)

    Scheers, B.; Bloemen, S.; Mühleisen, H.; Schellart, P.; van Elteren, A.; Kersten, M.; Groot, P. J.

    2018-04-01

    Coming high-cadence wide-field optical telescopes will image hundreds of thousands of sources per minute. Besides inspecting the near real-time data streams for transient and variability events, the accumulated data archive is a wealthy laboratory for making complementary scientific discoveries. The goal of this work is to optimise column-oriented database techniques to enable the construction of a full-source and light-curve database for large-scale surveys, that is accessible by the astronomical community. We adopted LOFAR's Transients Pipeline as the baseline and modified it to enable the processing of optical images that have much higher source densities. The pipeline adds new source lists to the archive database, while cross-matching them with the known cataloguedsources in order to build a full light-curve archive. We investigated several techniques of indexing and partitioning the largest tables, allowing for faster positional source look-ups in the cross matching algorithms. We monitored all query run times in long-term pipeline runs where we processed a subset of IPHAS data that have image source density peaks over 170,000 per field of view (500,000 deg-2). Our analysis demonstrates that horizontal table partitions of declination widths of one-degree control the query run times. Usage of an index strategy where the partitions are densely sorted according to source declination yields another improvement. Most queries run in sublinear time and a few (< 20%) run in linear time, because of dependencies on input source-list and result-set size. We observed that for this logical database partitioning schema the limiting cadence the pipeline achieved with processing IPHAS data is 25 s.

  16. Extending the search for neutrino point sources with IceCube above the horizon

    Energy Technology Data Exchange (ETDEWEB)

    IceCube Collaboration; Abbasi, R.

    2009-11-20

    Point source searches with the IceCube neutrino telescope have been restricted to one hemisphere, due to the exclusive selection of upward going events as a way of rejecting the atmospheric muon background. We show that the region above the horizon can be included by suppressing the background through energy-sensitive cuts. This approach improves the sensitivity above PeV energies, previously not accessible for declinations of more than a few degrees below the horizon due to the absorption of neutrinos in Earth. We present results based on data collected with 22 strings of IceCube, extending its field of view and energy reach for point source searches. No significant excess above the atmospheric background is observed in a sky scan and in tests of source candidates. Upper limits are reported, which for the first time cover point sources in the southern sky up to EeV energies.

  17. jSPyDB, an open source database-independent tool for data management

    Science.gov (United States)

    Pierro, Giuseppe Antonio; Cavallari, Francesca; Di Guida, Salvatore; Innocente, Vincenzo

    2011-12-01

    Nowadays, the number of commercial tools available for accessing Databases, built on Java or .Net, is increasing. However, many of these applications have several drawbacks: usually they are not open-source, they provide interfaces only with a specific kind of database, they are platform-dependent and very CPU and memory consuming. jSPyDB is a free web-based tool written using Python and Javascript. It relies on jQuery and python libraries, and is intended to provide a simple handler to different database technologies inside a local web browser. Such a tool, exploiting fast access libraries such as SQLAlchemy, is easy to install, and to configure. The design of this tool envisages three layers. The front-end client side in the local web browser communicates with a backend server. Only the server is able to connect to the different databases for the purposes of performing data definition and manipulation. The server makes the data available to the client, so that the user can display and handle them safely. Moreover, thanks to jQuery libraries, this tool supports export of data in different formats, such as XML and JSON. Finally, by using a set of pre-defined functions, users are allowed to create their customized views for a better data visualization. In this way, we optimize the performance of database servers by avoiding short connections and concurrent sessions. In addition, security is enforced since we do not provide users the possibility to directly execute any SQL statement.

  18. jSPyDB, an open source database-independent tool for data management

    International Nuclear Information System (INIS)

    Pierro, Giuseppe Antonio; Cavallari, Francesca; Di Guida, Salvatore; Innocente, Vincenzo

    2011-01-01

    Nowadays, the number of commercial tools available for accessing Databases, built on Java or .Net, is increasing. However, many of these applications have several drawbacks: usually they are not open-source, they provide interfaces only with a specific kind of database, they are platform-dependent and very CPU and memory consuming. jSPyDB is a free web-based tool written using Python and Javascript. It relies on jQuery and python libraries, and is intended to provide a simple handler to different database technologies inside a local web browser. Such a tool, exploiting fast access libraries such as SQLAlchemy, is easy to install, and to configure. The design of this tool envisages three layers. The front-end client side in the local web browser communicates with a backend server. Only the server is able to connect to the different databases for the purposes of performing data definition and manipulation. The server makes the data available to the client, so that the user can display and handle them safely. Moreover, thanks to jQuery libraries, this tool supports export of data in different formats, such as XML and JSON. Finally, by using a set of pre-defined functions, users are allowed to create their customized views for a better data visualization. In this way, we optimize the performance of database servers by avoiding short connections and concurrent sessions. In addition, security is enforced since we do not provide users the possibility to directly execute any SQL statement.

  19. Point source identification in nonlinear advection–diffusion–reaction systems

    International Nuclear Information System (INIS)

    Mamonov, A V; Tsai, Y-H R

    2013-01-01

    We consider a problem of identification of point sources in time-dependent advection–diffusion systems with a nonlinear reaction term. The linear counterpart of the problem in question can be reduced to solving a system of nonlinear algebraic equations via the use of adjoint equations. We extend this approach by constructing an algorithm that solves the problem iteratively to account for the nonlinearity of the reaction term. We study the question of improving the quality of source identification by adding more measurements adaptively using the solution obtained previously with a smaller number of measurements. (paper)

  20. Detection of Point Sources on Two-Dimensional Images Based on Peaks

    Directory of Open Access Journals (Sweden)

    R. B. Barreiro

    2005-09-01

    Full Text Available This paper considers the detection of point sources in two-dimensional astronomical images. The detection scheme we propose is based on peak statistics. We discuss the example of the detection of far galaxies in cosmic microwave background experiments throughout the paper, although the method we present is totally general and can be used in many other fields of data analysis. We consider sources with a Gaussian profile—that is, a fair approximation of the profile of a point source convolved with the detector beam in microwave experiments—on a background modeled by a homogeneous and isotropic Gaussian random field characterized by a scale-free power spectrum. Point sources are enhanced with respect to the background by means of linear filters. After filtering, we identify local maxima and apply our detection scheme, a Neyman-Pearson detector that defines our region of acceptance based on the a priori pdf of the sources and the ratio of number densities. We study the different performances of some linear filters that have been used in this context in the literature: the Mexican hat wavelet, the matched filter, and the scale-adaptive filter. We consider as well an extension to two dimensions of the biparametric scale-adaptive filter (BSAF. The BSAF depends on two parameters which are determined by maximizing the number density of real detections while fixing the number density of spurious detections. For our detection criterion the BSAF outperforms the other filters in the interesting case of white noise.

  1. Assessment of the impact of point source pollution from the ...

    African Journals Online (AJOL)

    Assessment of the impact of point source pollution from the Keiskammahoek Sewage ... Water SA. Journal Home · ABOUT THIS JOURNAL · Advanced Search ... Also, significant pollution of the receiving Keiskamma River was indicated for ...

  2. Tackling non-point source water pollution in British Columbia : an action plan

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1999-03-01

    British Columbia`s approach to water quality management is discussed. The BC efforts include regulating `end of pipe` point discharges from industrial and municipal outfalls. The major remaining cause of water pollution is from non-point sources (NPS). NPS water pollution is caused by the release of pollutants from different and diffuse sources, mostly unregulated and associated with urbanization, agriculture and other forms of land development. The importance of dealing with such problems on an immediate basis to avoid a decline in water quality in the province is emphasized. Major sources of water pollution in British Columbia include: land development, agriculture, storm water runoff, onsite sewage systems, forestry, atmospheric deposition, and marine activities. 3 tabs.

  3. Guaranteed Unresolved Point Source Emission and the Gamma-ray Background

    International Nuclear Information System (INIS)

    Pavlidou, Vasiliki; Siegal-Gaskins, Jennifer M.; Brown, Carolyn; Fields, Brian D.; Olinto, Angela V.

    2007-01-01

    The large majority of EGRET point sources remain without an identified low-energy counterpart, and a large fraction of these sources are most likely extragalactic. Whatever the nature of the extragalactic EGRET unidentified sources, faint unresolved objects of the same class must have a contribution to the diffuse extragalactic gamma-ray background (EGRB). Understanding this component of the EGRB, along with other guaranteed contributions from known sources (blazars and normal galaxies), is essential if we are to use this emission to constrain exotic high-energy physics. Here, we follow an empirical approach to estimate whether the contribution of unresolved unidentified sources to the EGRB is likely to be important. Additionally, we discuss how upcoming GLAST observations of EGRET unidentified sources, their fainter counterparts, and the Galactic and extragalactic diffuse backgrounds, will shed light on the nature of the EGRET unidentified sources even without any positional association of such sources with low-energy counterparts

  4. Evaluation of the Agricultural Non-point Source Pollution in Chongqing Based on PSR Model

    Institute of Scientific and Technical Information of China (English)

    Hanwen; ZHANG; Xinli; MOU; Hui; XIE; Hong; LU; Xingyun; YAN

    2014-01-01

    Through a series of exploration based on PSR framework model,for the purpose of building a suitable Chongqing agricultural nonpoint source pollution evaluation index system model framework,combined with the presence of Chongqing specific agro-environmental issues,we build a agricultural non-point source pollution assessment index system,and then study the agricultural system pressure,agro-environmental status and human response in total 3 major categories,develope an agricultural non-point source pollution evaluation index consisting of 3 criteria indicators and 19 indicators. As can be seen from the analysis,pressures and responses tend to increase and decrease linearly,state and complex have large fluctuations,and their fluctuations are similar mainly due to the elimination of pressures and impact,increasing the impact for agricultural non-point source pollution.

  5. A modified likelihood-method to search for point-sources in the diffuse astrophysical neutrino-flux in IceCube

    Energy Technology Data Exchange (ETDEWEB)

    Reimann, Rene; Haack, Christian; Leuermann, Martin; Raedel, Leif; Schoenen, Sebastian; Schimp, Michael; Wiebusch, Christopher [III. Physikalisches Institut, RWTH Aachen (Germany); Collaboration: IceCube-Collaboration

    2015-07-01

    IceCube, a cubic-kilometer sized neutrino detector at the geographical South Pole, has recently measured a flux of high-energy astrophysical neutrinos. Although this flux has now been observed in multiple analyses, no point sources or source classes could be identified yet. Standard point source searches test many points in the sky for a point source of astrophysical neutrinos individually and therefore produce many trials. Our approach is to additionally use the measured diffuse spectrum to constrain the number of possible point sources and their properties. Initial studies of the method performance are shown.

  6. Populating a Control Point Database: A cooperative effort between the USGS, Grand Canyon Monitoring and Research Center and the Grand Canyon Youth Organization

    Science.gov (United States)

    Brown, K. M.; Fritzinger, C.; Wharton, E.

    2004-12-01

    The Grand Canyon Monitoring and Research Center measures the effects of Glen Canyon Dam operations on the resources along the Colorado River from Glen Canyon Dam to Lake Mead in support of the Grand Canyon Adaptive Management Program. Control points are integral for geo-referencing the myriad of data collected in the Grand Canyon including aerial photography, topographic and bathymetric data used for classification and change-detection analysis of physical, biologic and cultural resources. The survey department has compiled a list of 870 control points installed by various organizations needing to establish a consistent reference for data collected at field sites along the 240 mile stretch of Colorado River in the Grand Canyon. This list is the foundation for the Control Point Database established primarily for researchers, to locate control points and independently geo-reference collected field data. The database has the potential to be a valuable mapping tool for assisting researchers to easily locate a control point and reduce the occurrance of unknowingly installing new control points within close proximity of an existing control point. The database is missing photographs and accurate site description information. Current site descriptions do not accurately define the location of the point but refer to the project that used the point, or some other interesting fact associated with the point. The Grand Canyon Monitoring and Research Center (GCMRC) resolved this problem by turning the data collection effort into an educational exercise for the participants of the Grand Canyon Youth organization. Grand Canyon Youth is a non-profit organization providing experiential education for middle and high school aged youth. GCMRC and the Grand Canyon Youth formed a partnership where GCMRC provided the logistical support, equipment, and training to conduct the field work, and the Grand Canyon Youth provided the time and personnel to complete the field work. Two data

  7. Nonpoint and Point Sources of Nitrogen in Major Watersheds of the United States

    Science.gov (United States)

    Puckett, Larry J.

    1994-01-01

    Estimates of nonpoint and point sources of nitrogen were made for 107 watersheds located in the U.S. Geological Survey's National Water-Quality Assessment Program study units throughout the conterminous United States. The proportions of nitrogen originating from fertilizer, manure, atmospheric deposition, sewage, and industrial sources were found to vary with climate, hydrologic conditions, land use, population, and physiography. Fertilizer sources of nitrogen are proportionally greater in agricultural areas of the West and the Midwest than in other parts of the Nation. Animal manure contributes large proportions of nitrogen in the South and parts of the Northeast. Atmospheric deposition of nitrogen is generally greatest in areas of greatest precipitation, such as the Northeast. Point sources (sewage and industrial) generally are predominant in watersheds near cities, where they may account for large proportions of the nitrogen in streams. The transport of nitrogen in streams increases as amounts of precipitation and runoff increase and is greatest in the Northeastern United States. Because no single nonpoint nitrogen source is dominant everywhere, approaches to control nitrogen must vary throughout the Nation. Watershed-based approaches to understanding nonpoint and point sources of contamination, as used by the National Water-Quality Assessment Program, will aid water-quality and environmental managers to devise methods to reduce nitrogen pollution.

  8. Specialist Bibliographic Databases.

    Science.gov (United States)

    Gasparyan, Armen Yuri; Yessirkepov, Marlen; Voronov, Alexander A; Trukhachev, Vladimir I; Kostyukova, Elena I; Gerasimov, Alexey N; Kitas, George D

    2016-05-01

    Specialist bibliographic databases offer essential online tools for researchers and authors who work on specific subjects and perform comprehensive and systematic syntheses of evidence. This article presents examples of the established specialist databases, which may be of interest to those engaged in multidisciplinary science communication. Access to most specialist databases is through subscription schemes and membership in professional associations. Several aggregators of information and database vendors, such as EBSCOhost and ProQuest, facilitate advanced searches supported by specialist keyword thesauri. Searches of items through specialist databases are complementary to those through multidisciplinary research platforms, such as PubMed, Web of Science, and Google Scholar. Familiarizing with the functional characteristics of biomedical and nonbiomedical bibliographic search tools is mandatory for researchers, authors, editors, and publishers. The database users are offered updates of the indexed journal lists, abstracts, author profiles, and links to other metadata. Editors and publishers may find particularly useful source selection criteria and apply for coverage of their peer-reviewed journals and grey literature sources. These criteria are aimed at accepting relevant sources with established editorial policies and quality controls.

  9. Specialist Bibliographic Databases

    Science.gov (United States)

    2016-01-01

    Specialist bibliographic databases offer essential online tools for researchers and authors who work on specific subjects and perform comprehensive and systematic syntheses of evidence. This article presents examples of the established specialist databases, which may be of interest to those engaged in multidisciplinary science communication. Access to most specialist databases is through subscription schemes and membership in professional associations. Several aggregators of information and database vendors, such as EBSCOhost and ProQuest, facilitate advanced searches supported by specialist keyword thesauri. Searches of items through specialist databases are complementary to those through multidisciplinary research platforms, such as PubMed, Web of Science, and Google Scholar. Familiarizing with the functional characteristics of biomedical and nonbiomedical bibliographic search tools is mandatory for researchers, authors, editors, and publishers. The database users are offered updates of the indexed journal lists, abstracts, author profiles, and links to other metadata. Editors and publishers may find particularly useful source selection criteria and apply for coverage of their peer-reviewed journals and grey literature sources. These criteria are aimed at accepting relevant sources with established editorial policies and quality controls. PMID:27134485

  10. Data Sources for Trait Databases: Comparing the Phenomic Content of Monographs and Evolutionary Matrices.

    Science.gov (United States)

    Dececchi, T Alex; Mabee, Paula M; Blackburn, David C

    2016-01-01

    Databases of organismal traits that aggregate information from one or multiple sources can be leveraged for large-scale analyses in biology. Yet the differences among these data streams and how well they capture trait diversity have never been explored. We present the first analysis of the differences between phenotypes captured in free text of descriptive publications ('monographs') and those used in phylogenetic analyses ('matrices'). We focus our analysis on osteological phenotypes of the limbs of four extinct vertebrate taxa critical to our understanding of the fin-to-limb transition. We find that there is low overlap between the anatomical entities used in these two sources of phenotype data, indicating that phenotypes represented in matrices are not simply a subset of those found in monographic descriptions. Perhaps as expected, compared to characters found in matrices, phenotypes in monographs tend to emphasize descriptive and positional morphology, be somewhat more complex, and relate to fewer additional taxa. While based on a small set of focal taxa, these qualitative and quantitative data suggest that either source of phenotypes alone will result in incomplete knowledge of variation for a given taxon. As a broader community develops to use and expand databases characterizing organismal trait diversity, it is important to recognize the limitations of the data sources and develop strategies to more fully characterize variation both within species and across the tree of life.

  11. Point-of-Care Healthcare Databases Are an Overall Asset to Clinicians, but Different Databases May Vary in Usefulness Based on Personal Preferences. A Review of: Chan, R. & Stieda, V. (2011). Evaluation of three point-of-care healthcare databases: BMJ Point-of-Care, Clin-eguide and Nursing Reference Centre. Health and Information Libraries Journal, 28(1), 50-58. doi: 10.1111/j.1471-1842.2010.00920.x

    OpenAIRE

    Carol D. Howe

    2011-01-01

    Objective – To evaluate the usefulness of three point-of-care healthcare databases (BMJ Point-of-Care, Clin-eguide, and Nursing Reference Centre) in clinical practice.Design – A descriptive study analyzing questionnaire results.Setting – Hospitals within Alberta, Canada’s two largest health regions (at the time of this study), with a third health region submitting a small number of responses.Subjects – A total of 46 Alberta hospital personnel answered the questionnaire, including 19 clinician...

  12. Refactoring databases evolutionary database design

    CERN Document Server

    Ambler, Scott W

    2006-01-01

    Refactoring has proven its value in a wide range of development projects–helping software professionals improve system designs, maintainability, extensibility, and performance. Now, for the first time, leading agile methodologist Scott Ambler and renowned consultant Pramodkumar Sadalage introduce powerful refactoring techniques specifically designed for database systems. Ambler and Sadalage demonstrate how small changes to table structures, data, stored procedures, and triggers can significantly enhance virtually any database design–without changing semantics. You’ll learn how to evolve database schemas in step with source code–and become far more effective in projects relying on iterative, agile methodologies. This comprehensive guide and reference helps you overcome the practical obstacles to refactoring real-world databases by covering every fundamental concept underlying database refactoring. Using start-to-finish examples, the authors walk you through refactoring simple standalone databas...

  13. Effect of point source and heterogeneity on the propagation of ...

    African Journals Online (AJOL)

    user

    propagation of Love waves due to point source in a homogeneous layer overlying a ...... The dispersion equation of SH waves will be obtained by equating to zero the ..... He was Awarded Atomic Energy Fellowship by the Government of India.

  14. Spatiotemporal patterns of non-point source nitrogen loss in an agricultural catchment

    Directory of Open Access Journals (Sweden)

    Jian-feng Xu

    2016-04-01

    Full Text Available Non-point source nitrogen loss poses a risk to sustainable aquatic ecosystems. However, non-point sources, as well as impaired river segments with high nitrogen concentrations, are difficult to monitor and regulate because of their diffusive nature, budget constraints, and resource deficiencies. For the purpose of catchment management, the Bayesian maximum entropy approach and spatial regression models have been used to explore the spatiotemporal patterns of non-point source nitrogen loss. In this study, a total of 18 sampling sites were selected along the river network in the Hujiashan Catchment. Over the time period of 2008–2012, water samples were collected 116 times at each site and analyzed for non-point source nitrogen loss. The morphometric variables and soil drainage of different land cover types were studied and considered potential factors affecting nitrogen loss. The results revealed that, compared with the approach using the Euclidean distance, the Bayesian maximum entropy approach using the river distance led to an appreciable 10.1% reduction in the estimation error, and more than 53.3% and 44.7% of the river network in the dry and wet seasons, respectively, had a probability of non-point source nitrogen impairment. The proportion of the impaired river segments exhibited an overall decreasing trend in the study catchment from 2008 to 2012, and the reduction in the wet seasons was greater than that in the dry seasons. High nitrogen concentrations were primarily found in the downstream reaches and river segments close to the residential lands. Croplands and residential lands were the dominant factors affecting non-point source nitrogen loss, and explained up to 70.7% of total nitrogen in the dry seasons and 54.7% in the wet seasons. A thorough understanding of the location of impaired river segments and the dominant factors affecting total nitrogen concentration would have considerable importance for catchment management.

  15. Screening the Medicines for Malaria Venture Pathogen Box across Multiple Pathogens Reclassifies Starting Points for Open-Source Drug Discovery.

    Science.gov (United States)

    Duffy, Sandra; Sykes, Melissa L; Jones, Amy J; Shelper, Todd B; Simpson, Moana; Lang, Rebecca; Poulsen, Sally-Ann; Sleebs, Brad E; Avery, Vicky M

    2017-09-01

    Open-access drug discovery provides a substantial resource for diseases primarily affecting the poor and disadvantaged. The open-access Pathogen Box collection is comprised of compounds with demonstrated biological activity against specific pathogenic organisms. The supply of this resource by the Medicines for Malaria Venture has the potential to provide new chemical starting points for a number of tropical and neglected diseases, through repurposing of these compounds for use in drug discovery campaigns for these additional pathogens. We tested the Pathogen Box against kinetoplastid parasites and malaria life cycle stages in vitro Consequently, chemical starting points for malaria, human African trypanosomiasis, Chagas disease, and leishmaniasis drug discovery efforts have been identified. Inclusive of this in vitro biological evaluation, outcomes from extensive literature reviews and database searches are provided. This information encompasses commercial availability, literature reference citations, other aliases and ChEMBL number with associated biological activity, where available. The release of this new data for the Pathogen Box collection into the public domain will aid the open-source model of drug discovery. Importantly, this will provide novel chemical starting points for drug discovery and target identification in tropical disease research. Copyright © 2017 Duffy et al.

  16. Developing an Inhouse Database from Online Sources.

    Science.gov (United States)

    Smith-Cohen, Deborah

    1993-01-01

    Describes the development of an in-house bibliographic database by the U.S. Army Corp of Engineers Cold Regions Research and Engineering Laboratory on arctic wetlands research. Topics discussed include planning; identifying relevant search terms and commercial online databases; downloading citations; criteria for software selection; management…

  17. Database system selection for marketing strategies support in information systems

    Directory of Open Access Journals (Sweden)

    František Dařena

    2007-01-01

    Full Text Available In today’s dynamically changing environment marketing has a significant role. Creating successful marketing strategies requires large amount of high quality information of various kinds and data types. A powerful database management system is a necessary condition for marketing strategies creation support. The paper briefly describes the field of marketing strategies and specifies the features that should be provided by database systems in connection with these strategies support. Major commercial (Oracle, DB2, MS SQL, Sybase and open-source (PostgreSQL, MySQL, Firebird databases are than examined from the point of view of accordance with these characteristics and their comparison in made. The results are useful for making the decision before acquisition of a database system during information system’s hardware architecture specification.

  18. IceCube point source searches using through-going muon tracks

    Energy Technology Data Exchange (ETDEWEB)

    Coenders, Stefan [TU Muenchen, Physik-Department, Excellence Cluster Universe, Boltzmannstr. 2, 85748 Garching (Germany); Collaboration: IceCube-Collaboration

    2015-07-01

    The IceCube neutrino observatory located at the South Pole is the current largest neutrino telescope. Using through-going muon tracks, IceCube records approximately 130,000 events per year with reconstruction accuracy as low as 0.7 deg for energies of 10 TeV. Having analysed an integrated time-scale of 4 years, no sources of neutrinos have yet been observed. This talk deals with the current progress in point-source searches, adding another two years of data recorded in the years 2012 and 2013. In a combined search with starting events, sources of hard and soft spectra with- and with-out cut-offs are characterised.

  19. Broadband integrated mid infrared light sources as enabling technology for point of care mid-infrared spectroscopy

    Science.gov (United States)

    2017-08-20

    AFRL-AFOSR-JP-TR-2017-0061 Broadband integrated mid-infrared light sources as enabling technology for point-of-care mid- infrared spectroscopy Alex...mid-infrared light sources as enabling technology for point-of-care mid-infrared spectroscopy 5a.  CONTRACT NUMBER 5b.  GRANT NUMBER FA2386-16-1-4037...Broadband integrated mid-infrared light sources as enabling technology for point-of-care mid- infrared spectroscopy ” Date: 16th August 2017 Name

  20. Microsoft Office SharePoint Designer 2007 Step by Step

    CERN Document Server

    Coventry, Penelope

    2008-01-01

    The smart way to learn Office SharePoint Designer 2007-one step at a time! Work at your own pace through the easy numbered steps, practice files on CD, helpful hints, and troubleshooting tips to master the fundamentals of building customized SharePoint sites and applications. You'll learn how to work with Windows® SharePoint Services 3.0 and Office SharePoint Server 2007 to create Web pages complete with Cascading Style Sheets, Lists, Libraries, and customized Web parts. Then, make your site really work for you by adding data sources, including databases, XML data and Web services, and RSS fe

  1. Evaluation of spatial dependence of point spread function-based PET reconstruction using a traceable point-like 22Na source

    Directory of Open Access Journals (Sweden)

    Taisuke Murata

    2016-10-01

    Full Text Available Abstract Background The point spread function (PSF of positron emission tomography (PET depends on the position across the field of view (FOV. Reconstruction based on PSF improves spatial resolution and quantitative accuracy. The present study aimed to quantify the effects of PSF correction as a function of the position of a traceable point-like 22Na source over the FOV on two PET scanners with a different detector design. Methods We used Discovery 600 and Discovery 710 (GE Healthcare PET scanners and traceable point-like 22Na sources (<1 MBq with a spherical absorber design that assures uniform angular distribution of the emitted annihilation photons. The source was moved in three directions at intervals of 1 cm from the center towards the peripheral FOV using a three-dimensional (3D-positioning robot, and data were acquired over a period of 2 min per point. The PET data were reconstructed by filtered back projection (FBP, the ordered subset expectation maximization (OSEM, OSEM + PSF, and OSEM + PSF + time-of-flight (TOF. Full width at half maximum (FWHM was determined according to the NEMA method, and total counts in regions of interest (ROI for each reconstruction were quantified. Results The radial FWHM of FBP and OSEM increased towards the peripheral FOV, whereas PSF-based reconstruction recovered the FWHM at all points in the FOV of both scanners. The radial FWHM for PSF was 30–50 % lower than that of OSEM at the center of the FOV. The accuracy of PSF correction was independent of detector design. Quantitative values were stable across the FOV in all reconstruction methods. The effect of TOF on spatial resolution and quantitation accuracy was less noticeable. Conclusions The traceable 22Na point-like source allowed the evaluation of spatial resolution and quantitative accuracy across the FOV using different reconstruction methods and scanners. PSF-based reconstruction reduces dependence of the spatial resolution on the

  2. Diffusion of dust particles from a point-source above ground level

    International Nuclear Information System (INIS)

    Hassan, M.H.A.; Eltayeb, I.A.

    1998-10-01

    A pollutant of small particles is emitted by a point source at a height h above ground level in an atmosphere in which a uni-directional wind speed, U, is prevailing. The pollutant is subjected to diffusion in all directions in the presence of advection and settling due to gravity. The equation governing the concentration of the pollutant is studied with the wind speed and the different components of diffusion tensor are proportional to the distance above ground level and the source has a uniform strength. Adopting a Cartesian system of coordinates in which the x-axis lies along the direction of the wind velocity, the z-axis is vertically upwards and the y-axis completes the right-hand triad, the solution for the concentration c(x,y,z) is obtained in closed form. The relative importance of the components of diffusion along the three axes is discussed. It is found that for any plane y=constant (=A), c(x,y,z) is concentrated along a curve of ''extensive pollution''. In the plane A=0, the concentration decreases along the line of extensive pollution as we move away from the source. However, for planes A≅0, the line of extensive pollution possesses a point of accumulation, which lies at a nonzero value of x. As we move away from the plane A=0, the point of accumulation moves laterally away from the plane x=0 and towards the plane z=0. The presence of the point of accumulation is entirely due to the presence of lateral diffusion. (author)

  3. Simultaneous Determination of Source Wavelet and Velocity Profile Using Impulsive Point-Source Reflections from a Layered Fluid

    National Research Council Canada - National Science Library

    Bube, K; Lailly, P; Sacks, P; Santosa, F; Symes, W. W

    1987-01-01

    .... We show that a quasi-impulsive, isotropic point source may be recovered simultaneously with the velocity profile from reflection data over a layered fluid, in linear (perturbation) approximation...

  4. ARACHNID: A prototype object-oriented database tool for distributed systems

    Science.gov (United States)

    Younger, Herbert; Oreilly, John; Frogner, Bjorn

    1994-01-01

    This paper discusses the results of a Phase 2 SBIR project sponsored by NASA and performed by MIMD Systems, Inc. A major objective of this project was to develop specific concepts for improved performance in accessing large databases. An object-oriented and distributed approach was used for the general design, while a geographical decomposition was used as a specific solution. The resulting software framework is called ARACHNID. The Faint Source Catalog developed by NASA was the initial database testbed. This is a database of many giga-bytes, where an order of magnitude improvement in query speed is being sought. This database contains faint infrared point sources obtained from telescope measurements of the sky. A geographical decomposition of this database is an attractive approach to dividing it into pieces. Each piece can then be searched on individual processors with only a weak data linkage between the processors being required. As a further demonstration of the concepts implemented in ARACHNID, a tourist information system is discussed. This version of ARACHNID is the commercial result of the project. It is a distributed, networked, database application where speed, maintenance, and reliability are important considerations. This paper focuses on the design concepts and technologies that form the basis for ARACHNID.

  5. Radio identifications of IRAS point sources with b greater than 30 deg

    International Nuclear Information System (INIS)

    Condon, J.J.; Broderick, J.J.; Virginia Polytechnic Institute and State Univ., Blacksburg)

    1986-01-01

    The present radio identifications of IRAS point sources on the basis of Green Bank 1400 MHz survey maps notes that 365 hot IR sources are not detectable radio sources, and that nearly all cool high latitude IRAS sources are extragalactic. The fainter IR-source identifications encompass optically bright quasars, BL Lac objects, Seyfert galaxies, and elliptical galaxies. No IRAS sources could be identified with distant elliptical radio galaxies, so that although the radio and IR fluxes of most IRAS extragalactic sources are tightly correlated, complete samples of strong radio and IR sources are almost completely disjoint; no more than 1 percent of the IR sources are radio sources and less than 1 percent of the radio sources are IR ones. 35 references

  6. Miniature x-ray point source for alignment and calibration of x-ray optics

    International Nuclear Information System (INIS)

    Price, R.H.; Boyle, M.J.; Glaros, S.S.

    1977-01-01

    A miniature x-ray point source of high brightness similar to that of Rovinsky, et al. is described. One version of the x-ray source is used to align the x-ray optics on the Argus and Shiva laser systems. A second version is used to determine the spatial and spectral transmission functions of the x-ray optics. The spatial and spectral characteristics of the x-ray emission from the x-ray point source are described. The physical constraints including size, intensity and thermal limitations, and useful lifetime are discussed. The alignment and calibration techniques for various x-ray optics and detector combinations are described

  7. Strategies for satellite-based monitoring of CO2 from distributed area and point sources

    Science.gov (United States)

    Schwandner, Florian M.; Miller, Charles E.; Duren, Riley M.; Natraj, Vijay; Eldering, Annmarie; Gunson, Michael R.; Crisp, David

    2014-05-01

    Atmospheric CO2 budgets are controlled by the strengths, as well as the spatial and temporal variabilities of CO2 sources and sinks. Natural CO2 sources and sinks are dominated by the vast areas of the oceans and the terrestrial biosphere. In contrast, anthropogenic and geogenic CO2 sources are dominated by distributed area and point sources, which may constitute as much as 70% of anthropogenic (e.g., Duren & Miller, 2012), and over 80% of geogenic emissions (Burton et al., 2013). Comprehensive assessments of CO2 budgets necessitate robust and highly accurate satellite remote sensing strategies that address the competing and often conflicting requirements for sampling over disparate space and time scales. Spatial variability: The spatial distribution of anthropogenic sources is dominated by patterns of production, storage, transport and use. In contrast, geogenic variability is almost entirely controlled by endogenic geological processes, except where surface gas permeability is modulated by soil moisture. Satellite remote sensing solutions will thus have to vary greatly in spatial coverage and resolution to address distributed area sources and point sources alike. Temporal variability: While biogenic sources are dominated by diurnal and seasonal patterns, anthropogenic sources fluctuate over a greater variety of time scales from diurnal, weekly and seasonal cycles, driven by both economic and climatic factors. Geogenic sources typically vary in time scales of days to months (geogenic sources sensu stricto are not fossil fuels but volcanoes, hydrothermal and metamorphic sources). Current ground-based monitoring networks for anthropogenic and geogenic sources record data on minute- to weekly temporal scales. Satellite remote sensing solutions would have to capture temporal variability through revisit frequency or point-and-stare strategies. Space-based remote sensing offers the potential of global coverage by a single sensor. However, no single combination of orbit

  8. Search for point-like sources using the diffuse astrophysical muon-neutrino flux in IceCube

    Energy Technology Data Exchange (ETDEWEB)

    Reimann, Rene; Haack, Christian; Raedel, Leif; Schoenen, Sebastian; Schumacher, Lisa; Wiebusch, Christopher [III. Physikalisches Institut B, RWTH Aachen (Germany); Collaboration: IceCube-Collaboration

    2016-07-01

    IceCube, a cubic-kilometer sized neutrino detector at the geographic South Pole, has recently confirmed a flux of high-energy astrophysical neutrinos in the track-like muon channel. Although this muon-neutrino flux has now been observed with high significance, no point sources or source classes could be identified yet with these well pointing events. We present a search for point-like sources based on a six year sample of upgoing muon-neutrinos with very low background contamination. To improve the sensitivity, the standard likelihood approach has been modified to focus on the properties of the measured astrophysical muon-neutrino flux.

  9. PSFGAN: a generative adversarial network system for separating quasar point sources and host galaxy light

    Science.gov (United States)

    Stark, Dominic; Launet, Barthelemy; Schawinski, Kevin; Zhang, Ce; Koss, Michael; Turp, M. Dennis; Sartori, Lia F.; Zhang, Hantian; Chen, Yiru; Weigel, Anna K.

    2018-06-01

    The study of unobscured active galactic nuclei (AGN) and quasars depends on the reliable decomposition of the light from the AGN point source and the extended host galaxy light. The problem is typically approached using parametric fitting routines using separate models for the host galaxy and the point spread function (PSF). We present a new approach using a Generative Adversarial Network (GAN) trained on galaxy images. We test the method using Sloan Digital Sky Survey r-band images with artificial AGN point sources added that are then removed using the GAN and with parametric methods using GALFIT. When the AGN point source is more than twice as bright as the host galaxy, we find that our method, PSFGAN, can recover point source and host galaxy magnitudes with smaller systematic error and a lower average scatter (49 per cent). PSFGAN is more tolerant to poor knowledge of the PSF than parametric methods. Our tests show that PSFGAN is robust against a broadening in the PSF width of ± 50 per cent if it is trained on multiple PSFs. We demonstrate that while a matched training set does improve performance, we can still subtract point sources using a PSFGAN trained on non-astronomical images. While initial training is computationally expensive, evaluating PSFGAN on data is more than 40 times faster than GALFIT fitting two components. Finally, PSFGAN is more robust and easy to use than parametric methods as it requires no input parameters.

  10. Is a wind turbine a point source? (L).

    Science.gov (United States)

    Makarewicz, Rufin

    2011-02-01

    Measurements show that practically all noise of wind turbine noise is produced by turbine blades, sometimes a few tens of meters long, despite that the model of a point source located at the hub height is commonly used. The plane of rotating blades is the critical location of the receiver because the distances to the blades are the shortest. It is shown that such location requires certain condition to be met. The model is valid far away from the wind turbine as well.

  11. Performance of Point and Range Queries for In-memory Databases using Radix Trees on GPUs

    Energy Technology Data Exchange (ETDEWEB)

    Alam, Maksudul [ORNL; Yoginath, Srikanth B [ORNL; Perumalla, Kalyan S [ORNL

    2016-01-01

    In in-memory database systems augmented by hardware accelerators, accelerating the index searching operations can greatly increase the runtime performance of database queries. Recently, adaptive radix trees (ART) have been shown to provide very fast index search implementation on the CPU. Here, we focus on an accelerator-based implementation of ART. We present a detailed performance study of our GPU-based adaptive radix tree (GRT) implementation over a variety of key distributions, synthetic benchmarks, and actual keys from music and book data sets. The performance is also compared with other index-searching schemes on the GPU. GRT on modern GPUs achieves some of the highest rates of index searches reported in the literature. For point queries, a throughput of up to 106 million and 130 million lookups per second is achieved for sparse and dense keys, respectively. For range queries, GRT yields 600 million and 1000 million lookups per second for sparse and dense keys, respectively, on a large dataset of 64 million 32-bit keys.

  12. Double point source W-phase inversion: Real-time implementation and automated model selection

    Science.gov (United States)

    Nealy, Jennifer; Hayes, Gavin

    2015-01-01

    Rapid and accurate characterization of an earthquake source is an extremely important and ever evolving field of research. Within this field, source inversion of the W-phase has recently been shown to be an effective technique, which can be efficiently implemented in real-time. An extension to the W-phase source inversion is presented in which two point sources are derived to better characterize complex earthquakes. A single source inversion followed by a double point source inversion with centroid locations fixed at the single source solution location can be efficiently run as part of earthquake monitoring network operational procedures. In order to determine the most appropriate solution, i.e., whether an earthquake is most appropriately described by a single source or a double source, an Akaike information criterion (AIC) test is performed. Analyses of all earthquakes of magnitude 7.5 and greater occurring since January 2000 were performed with extended analyses of the September 29, 2009 magnitude 8.1 Samoa earthquake and the April 19, 2014 magnitude 7.5 Papua New Guinea earthquake. The AIC test is shown to be able to accurately select the most appropriate model and the selected W-phase inversion is shown to yield reliable solutions that match published analyses of the same events.

  13. Temperature Effects of Point Sources, Riparian Shading, and Dam Operations on the Willamette River, Oregon

    Science.gov (United States)

    Rounds, Stewart A.

    2007-01-01

    Water temperature is an important factor influencing the migration, rearing, and spawning of several important fish species in rivers of the Pacific Northwest. To protect these fish populations and to fulfill its responsibilities under the Federal Clean Water Act, the Oregon Department of Environmental Quality set a water temperature Total Maximum Daily Load (TMDL) in 2006 for the Willamette River and the lower reaches of its largest tributaries in northwestern Oregon. As a result, the thermal discharges of the largest point sources of heat to the Willamette River now are limited at certain times of the year, riparian vegetation has been targeted for restoration, and upstream dams are recognized as important influences on downstream temperatures. Many of the prescribed point-source heat-load allocations are sufficiently restrictive that management agencies may need to expend considerable resources to meet those allocations. Trading heat allocations among point-source dischargers may be a more economical and efficient means of meeting the cumulative point-source temperature limits set by the TMDL. The cumulative nature of these limits, however, precludes simple one-to-one trades of heat from one point source to another; a more detailed spatial analysis is needed. In this investigation, the flow and temperature models that formed the basis of the Willamette temperature TMDL were used to determine a spatially indexed 'heating signature' for each of the modeled point sources, and those signatures then were combined into a user-friendly, spreadsheet-based screening tool. The Willamette River Point-Source Heat-Trading Tool allows the user to increase or decrease the heating signature of each source and thereby evaluate the effects of a wide range of potential point-source heat trades. The predictions of the Trading Tool were verified by running the Willamette flow and temperature models under four different trading scenarios, and the predictions typically were accurate

  14. NREL: U.S. Life Cycle Inventory Database - About the LCI Database Project

    Science.gov (United States)

    About the LCI Database Project The U.S. Life Cycle Inventory (LCI) Database is a publicly available database that allows users to objectively review and compare analysis results that are based on similar source of critically reviewed LCI data through its LCI Database Project. NREL's High-Performance

  15. Database Independent Migration of Objects into an Object-Relational Database

    CERN Document Server

    Ali, A; Munir, K; Waseem-Hassan, M; Willers, I

    2002-01-01

    CERN's (European Organization for Nuclear Research) WISDOM project [1] deals with the replication of data between homogeneous sources in a Wide Area Network (WAN) using the extensible Markup Language (XML). The last phase of the WISDOM (Wide-area, database Independent Serialization of Distributed Objects for data Migration) project [2], indicates the future directions for this work to be to incorporate heterogeneous sources as compared to homogeneous sources as described by [3]. This work will become essential for the CERN community once the need to transfer their legacy data to some other source, other then Objectivity [4], arises. Oracle 9i - an Object-Relational Database (including support for abstract data types, ADTs) appears to be a potential candidate for the physics event store in the CERN CMS experiment as suggested by [4] & [5]. Consequently this database has been selected for study. As a result of this work the HEP community will get a tool for migrating their data from Objectivity to Oracle9i.

  16. Lessons Learned from OMI Observations of Point Source SO2 Pollution

    Science.gov (United States)

    Krotkov, N.; Fioletov, V.; McLinden, Chris

    2011-01-01

    The Ozone Monitoring Instrument (OMI) on NASA Aura satellite makes global daily measurements of the total column of sulfur dioxide (SO2), a short-lived trace gas produced by fossil fuel combustion, smelting, and volcanoes. Although anthropogenic SO2 signals may not be detectable in a single OMI pixel, it is possible to see the source and determine its exact location by averaging a large number of individual measurements. We describe new techniques for spatial and temporal averaging that have been applied to the OMI SO2 data to determine the spatial distributions or "fingerprints" of SO2 burdens from top 100 pollution sources in North America. The technique requires averaging of several years of OMI daily measurements to observe SO2 pollution from typical anthropogenic sources. We found that the largest point sources of SO2 in the U.S. produce elevated SO2 values over a relatively small area - within 20-30 km radius. Therefore, one needs higher than OMI spatial resolution to monitor typical SO2 sources. TROPOMI instrument on the ESA Sentinel 5 precursor mission will have improved ground resolution (approximately 7 km at nadir), but is limited to once a day measurement. A pointable geostationary UVB spectrometer with variable spatial resolution and flexible sampling frequency could potentially achieve the goal of daily monitoring of SO2 point sources and resolve downwind plumes. This concept of taking the measurements at high frequency to enhance weak signals needs to be demonstrated with a GEOCAPE precursor mission before 2020, which will help formulating GEOCAPE measurement requirements.

  17. Identifying populations at risk from environmental contamination from point sources

    OpenAIRE

    Williams, F; Ogston, S

    2002-01-01

    Objectives: To compare methods for defining the population at risk from a point source of air pollution. A major challenge for environmental epidemiology lies in correctly identifying populations at risk from exposure to environmental pollutants. The complexity of today's environment makes it essential that the methods chosen are accurate and sensitive.

  18. Applicability of a desiccant dew-point cooling system independent of external water sources

    DEFF Research Database (Denmark)

    Bellemo, Lorenzo; Elmegaard, Brian; Kærn, Martin Ryhl

    2015-01-01

    The applicability of a technical solution for making desiccant cooling systems independent of external water sources is investigated. Water is produced by condensing the desorbed water vapour in a closed regeneration circuit. Desorbed water recovery is applied to a desiccant dew-point cooling...... system, which includes a desiccant wheel and a dew point cooler. The system is simulated during the summer period in the Mediterranean climate of Rome and it results completely independent of external water sources. The seasonal thermal COP drops 8% in comparison to the open regeneration circuit solution...

  19. An effective dose assessment technique with NORM added consumer products using skin-point source on computational human phantom

    International Nuclear Information System (INIS)

    Yoo, Do Hyeon; Shin, Wook-Geun; Lee, Hyun Cheol; Choi, Hyun Joon; Testa, Mauro; Lee, Jae Kook; Yeom, Yeon Soo; Kim, Chan Hyeong; Min, Chul Hee

    2016-01-01

    The aim of this study is to develop the assessment technique of the effective dose by calculating the organ equivalent dose with a Monte Carlo (MC) simulation and a computational human phantom for the naturally occurring radioactive material (NORM) added consumer products. In this study, we suggests the method determining the MC source term based on the skin-point source enabling the convenient and conservative modeling of the various type of the products. To validate the skin-point source method, the organ equivalent doses were compared with that by the product modeling source of the realistic shape for the pillow, waist supporter, sleeping mattress etc. Our results show that according to the source location, the organ equivalent doses were observed as the similar tendency for both source determining methods, however, it was observed that the annual effective dose with the skin-point source was conservative than that with the modeling source with the maximum 3.3 times higher dose. With the assumption of the gamma energy of 1 MeV and product activity of 1 Bq g"−"1, the annual effective doses of the pillow, waist supporter and sleeping mattress with skin-point source was 3.09E-16 Sv Bq"−"1 year"−"1, 1.45E-15 Sv Bq"−"1 year"−"1, and 2,82E-16 Sv Bq"−"1 year"−"1, respectively, while the product modeling source showed 9.22E-17 Sv Bq"−"1 year"−"1, 9.29E-16 Sv Bq"−"1 year"−"1, and 8.83E-17 Sv Bq"−"1 year"−"1, respectively. In conclusion, it was demonstrated in this study that the skin-point source method could be employed to efficiently evaluate the annual effective dose due to the usage of the NORM added consumer products. - Highlights: • We evaluate the exposure dose from the usage of NORM added consumer products. • We suggest the method determining the MC source term based on the skin-point source. • To validate the skin-point source, the organ equivalent doses were compared with that the modeling source. • The skin-point source could

  20. The database search problem: a question of rational decision making.

    Science.gov (United States)

    Gittelson, S; Biedermann, A; Bozza, S; Taroni, F

    2012-10-10

    This paper applies probability and decision theory in the graphical interface of an influence diagram to study the formal requirements of rationality which justify the individualization of a person found through a database search. The decision-theoretic part of the analysis studies the parameters that a rational decision maker would use to individualize the selected person. The modeling part (in the form of an influence diagram) clarifies the relationships between this decision and the ingredients that make up the database search problem, i.e., the results of the database search and the different pairs of propositions describing whether an individual is at the source of the crime stain. These analyses evaluate the desirability associated with the decision of 'individualizing' (and 'not individualizing'). They point out that this decision is a function of (i) the probability that the individual in question is, in fact, at the source of the crime stain (i.e., the state of nature), and (ii) the decision maker's preferences among the possible consequences of the decision (i.e., the decision maker's loss function). We discuss the relevance and argumentative implications of these insights with respect to recent comments in specialized literature, which suggest points of view that are opposed to the results of our study. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  1. Preparation of very small point sources for high resolution radiography

    International Nuclear Information System (INIS)

    Case, F.N.

    1976-01-01

    The need for very small point sources of high specific activity 192 Ir, 169 Yb, 170 Tm, and 60 Co in non-destructive testing has motivated the development of techniques for the fabrication of these sources. To prepare 192 Ir point sources for use in examination of tube sheet welds in LMFBR heat exchangers, 191 Ir enriched to greater than 90 percent was melted in a helium blanketed arc to form spheres as small as 0.38 mm in diameter. Methods were developed to form the roughly spherical shaped arc product into nearly symmetrical spheres that could be used for high resolution radiography. Similar methods were used for spherical shaped sources of 169 Yb and 170 Tm. The oxides were arc melted to form rough spheres followed by grinding to precise dimensions, neutron irradiation of the spheres at a flux of 2 to 3 x 10 15 nv, and use of enriched 168 Yb to provide the maximum specific activity. Cobalt-60 with a specific activity of greater than 1100 Ci/g was prepared by processing 59 Co that had been neutron irradiated to nearly complete burnup of the 59 Co target to produce 60 Co, 61 Ni, and 62 Ni. Ion exchange methods were used to separate the cobalt from the nickel. The cobalt was reduced to metal by plating either onto aluminum foil which was dissolved away from the cobalt plate, or by plating onto mercury to prepare amalgam that could be easily formed into a pellet of cobalt with exclusion of the mercury. Both methods are discussed

  2. Construction of a nasopharyngeal carcinoma 2D/MS repository with Open Source XML database--Xindice.

    Science.gov (United States)

    Li, Feng; Li, Maoyu; Xiao, Zhiqiang; Zhang, Pengfei; Li, Jianling; Chen, Zhuchu

    2006-01-11

    Many proteomics initiatives require integration of all information with uniformcriteria from collection of samples and data display to publication of experimental results. The integration and exchanging of these data of different formats and structure imposes a great challenge to us. The XML technology presents a promise in handling this task due to its simplicity and flexibility. Nasopharyngeal carcinoma (NPC) is one of the most common cancers in southern China and Southeast Asia, which has marked geographic and racial differences in incidence. Although there are some cancer proteome databases now, there is still no NPC proteome database. The raw NPC proteome experiment data were captured into one XML document with Human Proteome Markup Language (HUP-ML) editor and imported into native XML database Xindice. The 2D/MS repository of NPC proteome was constructed with Apache, PHP and Xindice to provide access to the database via Internet. On our website, two methods, keyword query and click query, were provided at the same time to access the entries of the NPC proteome database. Our 2D/MS repository can be used to share the raw NPC proteomics data that are generated from gel-based proteomics experiments. The database, as well as the PHP source codes for constructing users' own proteome repository, can be accessed at http://www.xyproteomics.org/.

  3. Point-of-Care Healthcare Databases Are an Overall Asset to Clinicians, but Different Databases May Vary in Usefulness Based on Personal Preferences. A Review of: Chan, R. & Stieda, V. (2011. Evaluation of three point-of-care healthcare databases: BMJ Point-of-Care, Clin-eguide and Nursing Reference Centre. Health and Information Libraries Journal, 28(1, 50-58. doi: 10.1111/j.1471-1842.2010.00920.x

    Directory of Open Access Journals (Sweden)

    Carol D. Howe

    2011-01-01

    Full Text Available Objective – To evaluate the usefulness of three point-of-care healthcare databases (BMJ Point-of-Care, Clin-eguide, and Nursing Reference Centre in clinical practice.Design – A descriptive study analyzing questionnaire results.Setting – Hospitals within Alberta, Canada’s two largest health regions (at the time of this study, with a third health region submitting a small number of responses.Subjects – A total of 46 Alberta hospital personnel answered the questionnaire, including 19 clinicians, 7 administrators, 6 nurses, 1 librarian, 1 preceptor, and “some” project coordinators. Subjects were chosen using a non-probability sampling method.Methods – The researchers developed an online questionnaire consisting of 17 questions and posted it on the University of Calgary’s Health Sciences Library and the Health Knowledge Network websites. The questions, in general, asked respondents how easy the databases were to search and use, whether the database content answered their clinical questions, and whether they would recommend the databases for future purchase. Most questions required a response for each of the three databases. The researchers collected quantitative data by using a Likert scale from 1 to 5, with 5 being the most positive answer and 1 being the most negative. They collected qualitative data by asking open-ended questions.Main Results – With regard to ease of searching, BMJ Point-of-Care (BMJ received the greatest number of responses (71% at level 5. A smaller number of respondents (56% rated Nursing Reference Centre (NRC at level 5. Clin-eguide received 59% of the responses at level 5, but it also received the greatest number of responses at the next highest level (level 4. Respondents rated all three databases similarly with regard to levels 1 and 2.Regarding how easy the resources were to learn, most respondents rated all three databases as easy to learn (BMJ, 77%; Clin-eguide, 72%; and NRC, 68%. Very few respondents

  4. Relationship Between Non-Point Source Pollution and Korean Green Factor

    Directory of Open Access Journals (Sweden)

    Seung Chul Lee

    2015-01-01

    Full Text Available In determining the relationship between the rational event mean concentration (REMC which is a volume-weighted mean of event mean concentrations (EMCs as a non-point source (NPS pollution indicator and the green factor (GF as a low impact development (LID land use planning indicator, we constructed at runoff database containing 1483 rainfall events collected from 107 different experimental catchments from 19 references in Korea. The collected data showed that EMCs were not correlated with storm factors whereas they showed significant differences according to the land use types. The calculated REMCs for BOD, COD, TSS, TN, and TP showed negative correlations with the GFs. However, even though the GFs of the agricultural area were concentrated in values of 80 like the green areas, the REMCs for TSS, TN, and TP were especially high. There were few differences in REMC runoff characteristics according to the GFs such as recreational facilities areas in suburbs and highways and trunk roads that connect to major roads between major cities. Except for those areas, the REMCs for BOD and COD were significantly related to the GFs. The REMCs for BOD and COD decreased when the rate of natural green area increased. On the other hand, some of the REMCs for TSS, TN, and TP were still high where the catchments encountered mixed land use patterns, especially public facility areas with bare ground and artificial grassland areas. The GF could therefore be used as a major planning indicator when establishing land use planning aimed at sustainable development with NPS management in urban areas if the weighted GF values will be improved.

  5. Diffusion from a point source in an urban atmosphere

    International Nuclear Information System (INIS)

    Essa, K.S.M.; El-Otaify, M.S.

    2005-01-01

    In the present paper, a model for the diffusion of material from a point source in an urban atmosphere is incorporated. The plume is assumed to have a well-defined edge at which the concentration falls to zero. The vertical wind shear is estimated using logarithmic law, by employing most of the available techniques of stability categories. The concentrations estimated from the model were compared favorably with the field observations of other investigators

  6. IceCube-Gen2 sensitivity improvement for steady neutrino point sources

    Energy Technology Data Exchange (ETDEWEB)

    Coenders, Stefan; Resconi, Elisa [TU Muenchen, Physik-Department, Excellence Cluster Universe, Boltzmannstr. 2, 85748 Garching (Germany); Collaboration: IceCube-Collaboration

    2015-07-01

    The observation of an astrophysical neutrino flux by high-energy events starting in IceCube strengthens the search for sources of astrophysical neutrinos. Identification of these sources requires good pointing at high statistics, mainly using muons created by charged-current muon neutrino interactions going through the IceCube detector. We report about preliminary studies of a possible high-energy extension IceCube-Gen2. Using a 6 times bigger detection volume, effective area as well as reconstruction accuracy will improve with respect to IceCube. Moreover, using (in-ice) active veto techniques will significantly improve the performance for Southern hemisphere events, where possible local candidate neutrino sources are located.

  7. HYDROLOGY AND SEDIMENT MODELING USING THE BASINS NON-POINT SOURCE MODEL

    Science.gov (United States)

    The Non-Point Source Model (Hydrologic Simulation Program-Fortran, or HSPF) within the EPA Office of Water's BASINS watershed modeling system was used to simulate streamflow and total suspended solids within Contentnea Creek, North Carolina, which is a tributary of the Neuse Rive...

  8. Nature of the Diffuse Source and Its Central Point-like Source in SNR 0509–67.5

    Energy Technology Data Exchange (ETDEWEB)

    Litke, Katrina C.; Chu, You-Hua; Holmes, Abigail; Santucci, Robert; Blindauer, Terrence; Gruendl, Robert A.; Ricker, Paul M. [Astronomy Department, University of Illinois, 1002 W. Green Street, Urbana, IL 61801 (United States); Li, Chuan-Jui [Academia Sinica Institute of Astronomy and Astrophysics, P.O. Box 23-141, Taipei 10617, Taiwan, R.O.C. (China); Pan, Kuo-Chuan [Departement Physik, Universität Basel, Klingelbergstrasse 82, CH-4056 Basel (Switzerland); Weisz, Daniel R., E-mail: kclitke@email.arizona.edu [Department of Astronomy, University of California, 501 Cambell Hall #3411, Berkeley, CA 94720-3411 (United States)

    2017-03-10

    We examine a diffuse emission region near the center of SNR 0509−67.5 to determine its nature. Within this diffuse region we observe a point-like source that is bright in the near-IR, but is not visible in the B and V bands. We consider an emission line observed at 6766 Å and the possibilities that it is Ly α , H α , and [O ii] λ 3727. We examine the spectral energy distribution (SED) of the source, comprised of Hubble Space Telescope B , V , I , J , and H bands in addition to Spitzer /IRAC 3.6, 4.5, 5.8, and 8 μ m bands. The peak of the SED is consistent with a background galaxy at z ≈ 0.8 ± 0.2 and a possible Balmer jump places the galaxy at z ≈ 0.9 ± 0.3. These SED considerations support the emission line’s identification as [O ii] λ 3727. We conclude that the diffuse source in SNR 0509−67.5 is a background galaxy at z ≈ 0.82. Furthermore, we identify the point-like source superposed near the center of the galaxy as its central bulge. Finally, we find no evidence for a surviving companion star, indicating a double-degenerate origin for SNR 0509−67.5.

  9. Nature of the Diffuse Source and Its Central Point-like Source in SNR 0509–67.5

    International Nuclear Information System (INIS)

    Litke, Katrina C.; Chu, You-Hua; Holmes, Abigail; Santucci, Robert; Blindauer, Terrence; Gruendl, Robert A.; Ricker, Paul M.; Li, Chuan-Jui; Pan, Kuo-Chuan; Weisz, Daniel R.

    2017-01-01

    We examine a diffuse emission region near the center of SNR 0509−67.5 to determine its nature. Within this diffuse region we observe a point-like source that is bright in the near-IR, but is not visible in the B and V bands. We consider an emission line observed at 6766 Å and the possibilities that it is Ly α , H α , and [O ii] λ 3727. We examine the spectral energy distribution (SED) of the source, comprised of Hubble Space Telescope B , V , I , J , and H bands in addition to Spitzer /IRAC 3.6, 4.5, 5.8, and 8 μ m bands. The peak of the SED is consistent with a background galaxy at z ≈ 0.8 ± 0.2 and a possible Balmer jump places the galaxy at z ≈ 0.9 ± 0.3. These SED considerations support the emission line’s identification as [O ii] λ 3727. We conclude that the diffuse source in SNR 0509−67.5 is a background galaxy at z ≈ 0.82. Furthermore, we identify the point-like source superposed near the center of the galaxy as its central bulge. Finally, we find no evidence for a surviving companion star, indicating a double-degenerate origin for SNR 0509−67.5.

  10. Active control on high-order coherence and statistic characterization on random phase fluctuation of two classical point sources.

    Science.gov (United States)

    Hong, Peilong; Li, Liming; Liu, Jianji; Zhang, Guoquan

    2016-03-29

    Young's double-slit or two-beam interference is of fundamental importance to understand various interference effects, in which the stationary phase difference between two beams plays the key role in the first-order coherence. Different from the case of first-order coherence, in the high-order optical coherence the statistic behavior of the optical phase will play the key role. In this article, by employing a fundamental interfering configuration with two classical point sources, we showed that the high- order optical coherence between two classical point sources can be actively designed by controlling the statistic behavior of the relative phase difference between two point sources. Synchronous position Nth-order subwavelength interference with an effective wavelength of λ/M was demonstrated, in which λ is the wavelength of point sources and M is an integer not larger than N. Interestingly, we found that the synchronous position Nth-order interference fringe fingerprints the statistic trace of random phase fluctuation of two classical point sources, therefore, it provides an effective way to characterize the statistic properties of phase fluctuation for incoherent light sources.

  11. Decreasing Computational Time for VBBinaryLensing by Point Source Approximation

    Science.gov (United States)

    Tirrell, Bethany M.; Visgaitis, Tiffany A.; Bozza, Valerio

    2018-01-01

    The gravitational lens of a binary system produces a magnification map that is more intricate than a single object lens. This map cannot be calculated analytically and one must rely on computational methods to resolve. There are generally two methods of computing the microlensed flux of a source. One is based on ray-shooting maps (Kayser, Refsdal, & Stabell 1986), while the other method is based on an application of Green’s theorem. This second method finds the area of an image by calculating a Riemann integral along the image contour. VBBinaryLensing is a C++ contour integration code developed by Valerio Bozza, which utilizes this method. The parameters at which the source object could be treated as a point source, or in other words, when the source is far enough from the caustic, was of interest to substantially decrease the computational time. The maximum and minimum values of the caustic curves produced, were examined to determine the boundaries for which this simplification could be made. The code was then run for a number of different maps, with separation values and accuracies ranging from 10-1 to 10-3, to test the theoretical model and determine a safe buffer for which minimal error could be made for the approximation. The determined buffer was 1.5+5q, with q being the mass ratio. The theoretical model and the calculated points worked for all combinations of the separation values and different accuracies except the map with accuracy and separation equal to 10-3 for y1 max. An alternative approach has to be found in order to accommodate a wider range of parameters.

  12. Database Organisation in a Web-Enabled Free and Open-Source Software (foss) Environment for Spatio-Temporal Landslide Modelling

    Science.gov (United States)

    Das, I.; Oberai, K.; Sarathi Roy, P.

    2012-07-01

    Landslides exhibit themselves in different mass movement processes and are considered among the most complex natural hazards occurring on the earth surface. Making landslide database available online via WWW (World Wide Web) promotes the spreading and reaching out of the landslide information to all the stakeholders. The aim of this research is to present a comprehensive database for generating landslide hazard scenario with the help of available historic records of landslides and geo-environmental factors and make them available over the Web using geospatial Free & Open Source Software (FOSS). FOSS reduces the cost of the project drastically as proprietary software's are very costly. Landslide data generated for the period 1982 to 2009 were compiled along the national highway road corridor in Indian Himalayas. All the geo-environmental datasets along with the landslide susceptibility map were served through WEBGIS client interface. Open source University of Minnesota (UMN) mapserver was used as GIS server software for developing web enabled landslide geospatial database. PHP/Mapscript server-side application serve as a front-end application and PostgreSQL with PostGIS extension serve as a backend application for the web enabled landslide spatio-temporal databases. This dynamic virtual visualization process through a web platform brings an insight into the understanding of the landslides and the resulting damage closer to the affected people and user community. The landslide susceptibility dataset is also made available as an Open Geospatial Consortium (OGC) Web Feature Service (WFS) which can be accessed through any OGC compliant open source or proprietary GIS Software.

  13. Reduction Assessment of Agricultural Non-Point Source Pollutant Loading

    OpenAIRE

    Fu, YiCheng; Zang, Wenbin; Zhang, Jian; Wang, Hongtao; Zhang, Chunling; Shi, Wanli

    2018-01-01

    NPS (Non-point source) pollution has become a key impact element to watershed environment at present. With the development of technology, application of models to control NPS pollution has become a very common practice for resource management and Pollutant reduction control in the watershed scale of China. The SWAT (Soil and Water Assessment Tool) model is a semi-conceptual model, which was put forward to estimate pollutant production & the influences on water quantity-quality under different...

  14. 75 FR 10438 - Effluent Limitations Guidelines and Standards for the Construction and Development Point Source...

    Science.gov (United States)

    2010-03-08

    ... Effluent Limitations Guidelines and Standards for the Construction and Development Point Source Category... technology-based Effluent Limitations Guidelines and New Source Performance Standards for the Construction... technology-based Effluent Limitations Guidelines and New Source Performance Standards for the Construction...

  15. DOE technology information management system database study report

    Energy Technology Data Exchange (ETDEWEB)

    Widing, M.A.; Blodgett, D.W.; Braun, M.D.; Jusko, M.J.; Keisler, J.M.; Love, R.J.; Robinson, G.L. [Argonne National Lab., IL (United States). Decision and Information Sciences Div.

    1994-11-01

    To support the missions of the US Department of Energy (DOE) Special Technologies Program, Argonne National Laboratory is defining the requirements for an automated software system that will search electronic databases on technology. This report examines the work done and results to date. Argonne studied existing commercial and government sources of technology databases in five general areas: on-line services, patent database sources, government sources, aerospace technology sources, and general technology sources. First, it conducted a preliminary investigation of these sources to obtain information on the content, cost, frequency of updates, and other aspects of their databases. The Laboratory then performed detailed examinations of at least one source in each area. On this basis, Argonne recommended which databases should be incorporated in DOE`s Technology Information Management System.

  16. Full Data of Yeast Interacting Proteins Database (Original Version) - Yeast Interacting Proteins Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Yeast Interacting Proteins Database Full Data of Yeast Interacting Proteins Database (Origin...al Version) Data detail Data name Full Data of Yeast Interacting Proteins Database (Original Version) DOI 10....18908/lsdba.nbdc00742-004 Description of data contents The entire data in the Yeast Interacting Proteins Database...eir interactions are required. Several sources including YPD (Yeast Proteome Database, Costanzo, M. C., Hoga...ematic name in the SGD (Saccharomyces Genome Database; http://www.yeastgenome.org /). Bait gene name The gen

  17. The Unicellular State as a Point Source in a Quantum Biological System

    Directory of Open Access Journals (Sweden)

    John S. Torday

    2016-05-01

    Full Text Available A point source is the central and most important point or place for any group of cohering phenomena. Evolutionary development presumes that biological processes are sequentially linked, but neither directed from, nor centralized within, any specific biologic structure or stage. However, such an epigenomic entity exists and its transforming effects can be understood through the obligatory recapitulation of all eukaryotic lifeforms through a zygotic unicellular phase. This requisite biological conjunction can now be properly assessed as the focal point of reconciliation between biology and quantum phenomena, illustrated by deconvoluting complex physiologic traits back to their unicellular origins.

  18. Calibrate the aerial surveying instrument by the limited surface source and the single point source that replace the unlimited surface source

    CERN Document Server

    Lu Cun Heng

    1999-01-01

    It is described that the calculating formula and surveying result is found on the basis of the stacking principle of gamma ray and the feature of hexagonal surface source when the limited surface source replaces the unlimited surface source to calibrate the aerial survey instrument on the ground, and that it is found in the light of the exchanged principle of the gamma ray when the single point source replaces the unlimited surface source to calibrate aerial surveying instrument in the air. Meanwhile through the theoretical analysis, the receiving rate of the crystal bottom and side surfaces is calculated when aerial surveying instrument receives gamma ray. The mathematical expression of the gamma ray decaying following height according to the Jinge function regularity is got. According to this regularity, the absorbing coefficient that air absorbs the gamma ray and the detective efficiency coefficient of the crystal is calculated based on the ground and air measuring value of the bottom surface receiving cou...

  19. Ambiguity of non-systematic chemical identifiers within and between small-molecule databases.

    Science.gov (United States)

    Akhondi, Saber A; Muresan, Sorel; Williams, Antony J; Kors, Jan A

    2015-01-01

    A wide range of chemical compound databases are currently available for pharmaceutical research. To retrieve compound information, including structures, researchers can query these chemical databases using non-systematic identifiers. These are source-dependent identifiers (e.g., brand names, generic names), which are usually assigned to the compound at the point of registration. The correctness of non-systematic identifiers (i.e., whether an identifier matches the associated structure) can only be assessed manually, which is cumbersome, but it is possible to automatically check their ambiguity (i.e., whether an identifier matches more than one structure). In this study we have quantified the ambiguity of non-systematic identifiers within and between eight widely used chemical databases. We also studied the effect of chemical structure standardization on reducing the ambiguity of non-systematic identifiers. The ambiguity of non-systematic identifiers within databases varied from 0.1 to 15.2 % (median 2.5 %). Standardization reduced the ambiguity only to a small extent for most databases. A wide range of ambiguity existed for non-systematic identifiers that are shared between databases (17.7-60.2 %, median of 40.3 %). Removing stereochemistry information provided the largest reduction in ambiguity across databases (median reduction 13.7 percentage points). Ambiguity of non-systematic identifiers within chemical databases is generally low, but ambiguity of non-systematic identifiers that are shared between databases, is high. Chemical structure standardization reduces the ambiguity to a limited extent. Our findings can help to improve database integration, curation, and maintenance.

  20. Effect of tissue inhomogeneity on dose distribution of point sources of low-energy electrons

    International Nuclear Information System (INIS)

    Kwok, C.S.; Bialobzyski, P.J.; Yu, S.K.; Prestwich, W.V.

    1990-01-01

    Perturbation in dose distributions of point sources of low-energy electrons at planar interfaces of cortical bone (CB) and red marrow (RM) was investigated experimentally and by Monte Carlo codes EGS and the TIGER series. Ultrathin LiF thermoluminescent dosimeters were used to measure the dose distributions of point sources of 204 Tl and 147 Pm in RM. When the point sources were at 12 mg/cm 2 from a planar interface of CB and RM equivalent plastics, dose enhancement ratios in RM averaged over the region 0--12 mg/cm 2 from the interface were measured to be 1.08±0.03 (SE) and 1.03±0.03 (SE) for 204 Tl and 147 Pm, respectively. The Monte Carlo codes predicted 1.05±0.02 and 1.01±0.02 for the two nuclides, respectively. However, EGS gave consistently 3% higher dose in the dose scoring region than the TIGER series when point sources of monoenergetic electrons up to 0.75 MeV energy were considered in the homogeneous RM situation or in the CB and RM heterogeneous situation. By means of the TIGER series, it was demonstrated that aluminum, which is normally assumed to be equivalent to CB in radiation dosimetry, leads to an overestimation of backscattering of low-energy electrons in soft tissue at a CB--soft-tissue interface by as much as a factor of 2

  1. Model for Semantically Rich Point Cloud Data

    Science.gov (United States)

    Poux, F.; Neuville, R.; Hallot, P.; Billen, R.

    2017-10-01

    This paper proposes an interoperable model for managing high dimensional point clouds while integrating semantics. Point clouds from sensors are a direct source of information physically describing a 3D state of the recorded environment. As such, they are an exhaustive representation of the real world at every scale: 3D reality-based spatial data. Their generation is increasingly fast but processing routines and data models lack of knowledge to reason from information extraction rather than interpretation. The enhanced smart point cloud developed model allows to bring intelligence to point clouds via 3 connected meta-models while linking available knowledge and classification procedures that permits semantic injection. Interoperability drives the model adaptation to potentially many applications through specialized domain ontologies. A first prototype is implemented in Python and PostgreSQL database and allows to combine semantic and spatial concepts for basic hybrid queries on different point clouds.

  2. MODEL FOR SEMANTICALLY RICH POINT CLOUD DATA

    Directory of Open Access Journals (Sweden)

    F. Poux

    2017-10-01

    Full Text Available This paper proposes an interoperable model for managing high dimensional point clouds while integrating semantics. Point clouds from sensors are a direct source of information physically describing a 3D state of the recorded environment. As such, they are an exhaustive representation of the real world at every scale: 3D reality-based spatial data. Their generation is increasingly fast but processing routines and data models lack of knowledge to reason from information extraction rather than interpretation. The enhanced smart point cloud developed model allows to bring intelligence to point clouds via 3 connected meta-models while linking available knowledge and classification procedures that permits semantic injection. Interoperability drives the model adaptation to potentially many applications through specialized domain ontologies. A first prototype is implemented in Python and PostgreSQL database and allows to combine semantic and spatial concepts for basic hybrid queries on different point clouds.

  3. PSFGAN: a generative adversarial network system for separating quasar point sources and host galaxy light

    Science.gov (United States)

    Stark, Dominic; Launet, Barthelemy; Schawinski, Kevin; Zhang, Ce; Koss, Michael; Turp, M. Dennis; Sartori, Lia F.; Zhang, Hantian; Chen, Yiru; Weigel, Anna K.

    2018-03-01

    The study of unobscured active galactic nuclei (AGN) and quasars depends on the reliable decomposition of the light from the AGN point source and the extended host galaxy light. The problem is typically approached using parametric fitting routines using separate models for the host galaxy and the point spread function (PSF). We present a new approach using a Generative Adversarial Network (GAN) trained on galaxy images. We test the method using Sloan Digital Sky Survey (SDSS) r-band images with artificial AGN point sources added which are then removed using the GAN and with parametric methods using GALFIT. When the AGN point source PS is more than twice as bright as the host galaxy, we find that our method, PSFGAN, can recover PS and host galaxy magnitudes with smaller systematic error and a lower average scatter (49%). PSFGAN is more tolerant to poor knowledge of the PSF than parametric methods. Our tests show that PSFGAN is robust against a broadening in the PSF width of ±50% if it is trained on multiple PSF's. We demonstrate that while a matched training set does improve performance, we can still subtract point sources using a PSFGAN trained on non-astronomical images. While initial training is computationally expensive, evaluating PSFGAN on data is more than 40 times faster than GALFIT fitting two components. Finally, PSFGAN it is more robust and easy to use than parametric methods as it requires no input parameters.

  4. Free and Open Source Options for Creating Database-Driven Subject Guides

    Directory of Open Access Journals (Sweden)

    Edward M. Corrado

    2008-03-01

    Full Text Available This article reviews available cost-effective options libraries have for updating and maintaining pathfinders such as subject guides and course pages. The paper discusses many of the available options, from the standpoint of a mid-sized academic library which is evaluating alternatives to static-HTML subject guides. Static HTML guides, while useful, have proven difficult and time-consuming to maintain. The article includes a discussion of open source database-driven solutions (such as SubjectsPlus, LibData, Research Guide, and Library Course Builder, Wikis, and social tagging sites like del.icio.us. This article discusses both the functionality and the relative strengths and weaknessess of each of these options.

  5. Construction of a nasopharyngeal carcinoma 2D/MS repository with Open Source XML Database – Xindice

    Directory of Open Access Journals (Sweden)

    Li Jianling

    2006-01-01

    Full Text Available Abstract Background Many proteomics initiatives require integration of all information with uniformcriteria from collection of samples and data display to publication of experimental results. The integration and exchanging of these data of different formats and structure imposes a great challenge to us. The XML technology presents a promise in handling this task due to its simplicity and flexibility. Nasopharyngeal carcinoma (NPC is one of the most common cancers in southern China and Southeast Asia, which has marked geographic and racial differences in incidence. Although there are some cancer proteome databases now, there is still no NPC proteome database. Results The raw NPC proteome experiment data were captured into one XML document with Human Proteome Markup Language (HUP-ML editor and imported into native XML database Xindice. The 2D/MS repository of NPC proteome was constructed with Apache, PHP and Xindice to provide access to the database via Internet. On our website, two methods, keyword query and click query, were provided at the same time to access the entries of the NPC proteome database. Conclusion Our 2D/MS repository can be used to share the raw NPC proteomics data that are generated from gel-based proteomics experiments. The database, as well as the PHP source codes for constructing users' own proteome repository, can be accessed at http://www.xyproteomics.org/.

  6. Global search tool for the Advanced Photon Source Integrated Relational Model of Installed Systems (IRMIS) database

    International Nuclear Information System (INIS)

    Quock, D.E.R.; Cianciarulo, M.B.

    2007-01-01

    The Integrated Relational Model of Installed Systems (IRMIS) is a relational database tool that has been implemented at the Advanced Photon Source to maintain an updated account of approximately 600 control system software applications, 400,000 process variables, and 30,000 control system hardware components. To effectively display this large amount of control system information to operators and engineers, IRMIS was initially built with nine Web-based viewers: Applications Organizing Index, IOC, PLC, Component Type, Installed Components, Network, Controls Spares, Process Variables, and Cables. However, since each viewer is designed to provide details from only one major category of the control system, the necessity for a one-stop global search tool for the entire database became apparent. The user requirements for extremely fast database search time and ease of navigation through search results led to the choice of Asynchronous JavaScript and XML (AJAX) technology in the implementation of the IRMIS global search tool. Unique features of the global search tool include a two-tier level of displayed search results, and a database data integrity validation and reporting mechanism.

  7. Simulation of ultrasonic surface waves with multi-Gaussian and point source beam models

    International Nuclear Information System (INIS)

    Zhao, Xinyu; Schmerr, Lester W. Jr.; Li, Xiongbing; Sedov, Alexander

    2014-01-01

    In the past decade, multi-Gaussian beam models have been developed to solve many complicated bulk wave propagation problems. However, to date those models have not been extended to simulate the generation of Rayleigh waves. Here we will combine Gaussian beams with an explicit high frequency expression for the Rayleigh wave Green function to produce a three-dimensional multi-Gaussian beam model for the fields radiated from an angle beam transducer mounted on a solid wedge. Simulation results obtained with this model are compared to those of a point source model. It is shown that the multi-Gaussian surface wave beam model agrees well with the point source model while being computationally much more efficient

  8. Time dependence of the field energy densities surrounding sources: Application to scalar mesons near point sources and to electromagnetic fields near molecules

    International Nuclear Information System (INIS)

    Persico, F.; Power, E.A.

    1987-01-01

    The time dependence of the dressing-undressing process, i.e., the acquiring or losing by a source of a boson field intensity and hence of a field energy density in its neighborhood, is considered by examining some simple soluble models. First, the loss of the virtual field is followed in time when a point source is suddenly decoupled from a neutral scalar meson field. Second, an initially bare point source acquires a virtual meson cloud as the coupling is switched on. The third example is that of an initially bare molecule interacting with the vacuum of the electromagnetic field to acquire a virtual photon cloud. In all three cases the dressing-undressing is shown to take place within an expanding sphere of radius r = ct centered at the source. At each point in space the energy density tends, for large times, to that of the ground state of the total system. Differences in the time dependence of the dressing between the massive scalar field and the massless electromagnetic field are discussed. The results are also briefly discussed in the light of Feinberg's ideas on the nature of half-dressed states in quantum field theory

  9. Fast and Accurate Rat Head Motion Tracking With Point Sources for Awake Brain PET.

    Science.gov (United States)

    Miranda, Alan; Staelens, Steven; Stroobants, Sigrid; Verhaeghe, Jeroen

    2017-07-01

    To avoid the confounding effects of anesthesia and immobilization stress in rat brain positron emission tomography (PET), motion tracking-based unrestrained awake rat brain imaging is being developed. In this paper, we propose a fast and accurate rat headmotion tracking method based on small PET point sources. PET point sources (3-4) attached to the rat's head are tracked in image space using 15-32-ms time frames. Our point source tracking (PST) method was validated using a manually moved microDerenzo phantom that was simultaneously tracked with an optical tracker (OT) for comparison. The PST method was further validated in three awake [ 18 F]FDG rat brain scans. Compared with the OT, the PST-based correction at the same frame rate (31.2 Hz) reduced the reconstructed FWHM by 0.39-0.66 mm for the different tested rod sizes of the microDerenzo phantom. The FWHM could be further reduced by another 0.07-0.13 mm when increasing the PST frame rate (66.7 Hz). Regional brain [ 18 F]FDG uptake in the motion corrected scan was strongly correlated ( ) with that of the anesthetized reference scan for all three cases ( ). The proposed PST method allowed excellent and reproducible motion correction in awake in vivo experiments. In addition, there is no need of specialized tracking equipment or additional calibrations to be performed, the point sources are practically imperceptible to the rat, and PST is ideally suitable for small bore scanners, where optical tracking might be challenging.

  10. Brasilia’s Database Administrators

    Directory of Open Access Journals (Sweden)

    Jane Adriana

    2016-06-01

    Full Text Available Database administration has gained an essential role in the management of new database technologies. Different data models are being created for supporting the enormous data volume, from the traditional relational database. These new models are called NoSQL (Not only SQL databases. The adoption of best practices and procedures, has become essential for the operation of database management systems. Thus, this paper investigates some of the techniques and tools used by database administrators. The study highlights features and particularities in databases within the area of Brasilia, the Capital of Brazil. The results point to which new technologies regarding database management are currently the most relevant, as well as the central issues in this area.

  11. Seasonal and spatial variation of diffuse (non-point) source zinc pollution in a historically metal mined river catchment, UK

    Energy Technology Data Exchange (ETDEWEB)

    Gozzard, E., E-mail: emgo@ceh.ac.uk [Hydrogeochemical Engineering Research and Outreach Group, School of Civil Engineering and Geosciences, Newcastle University, Newcastle upon Tyne NE1 7RU (United Kingdom); Mayes, W.M., E-mail: W.Mayes@hull.ac.uk [Hydrogeochemical Engineering Research and Outreach Group, School of Civil Engineering and Geosciences, Newcastle University, Newcastle upon Tyne NE1 7RU (United Kingdom); Potter, H.A.B., E-mail: hugh.potter@environment-agency.gov.uk [Environment Agency England and Wales, c/o Institute for Research on Environment and Sustainability, Newcastle University, Newcastle upon Tyne NE1 7RU (United Kingdom); Jarvis, A.P., E-mail: a.p.jarvis@ncl.ac.uk [Hydrogeochemical Engineering Research and Outreach Group, School of Civil Engineering and Geosciences, Newcastle University, Newcastle upon Tyne NE1 7RU (United Kingdom)

    2011-10-15

    Quantifying diffuse sources of pollution is becoming increasingly important when characterising river catchments in entirety - a prerequisite for environmental management. This study examines both low and high flow events, as well as spatial variability, in order to assess point and diffuse components of zinc pollution within the River West Allen catchment, which lies within the northern England lead-zinc Orefield. Zinc levels in the river are elevated under all flow regimes, and are of environmental concern. Diffuse components are of little importance at low flow, with point source mine water discharges dominating instream zinc concentration and load. During higher river flows 90% of the instream zinc load is attributed to diffuse sources, where inputs from resuspension of metal-rich sediments, and groundwater influx are likely to be more dominant. Remediating point mine water discharges should significantly improve water quality at lower flows, but contribution from diffuse sources will continue to elevate zinc flux at higher flows. - Highlights: > Zinc concentrations breach EU quality thresholds under all river flow conditions. > Contributions from point sources dominate instream zinc dynamics in low flow. > Contributions from diffuse sources dominate instream zinc dynamics in high flow. > Important diffuse sources include river-bed sediment resuspension and groundwater influx. > Diffuse sources would still create significant instream pollution, even with point source treatment. - Diffuse zinc sources are an important source of instream contamination to mine-impacted rivers under varying flow conditions.

  12. A Review of Stellar Abundance Databases and the Hypatia Catalog Database

    Science.gov (United States)

    Hinkel, Natalie Rose

    2018-01-01

    The astronomical community is interested in elements from lithium to thorium, from solar twins to peculiarities of stellar evolution, because they give insight into different regimes of star formation and evolution. However, while some trends between elements and other stellar or planetary properties are well known, many other trends are not as obvious and are a point of conflict. For example, stars that host giant planets are found to be consistently enriched in iron, but the same cannot be definitively said for any other element. Therefore, it is time to take advantage of large stellar abundance databases in order to better understand not only the large-scale patterns, but also the more subtle, small-scale trends within the data.In this overview to the special session, I will present a review of large stellar abundance databases that are both currently available (i.e. RAVE, APOGEE) and those that will soon be online (i.e. Gaia-ESO, GALAH). Additionally, I will discuss the Hypatia Catalog Database (www.hypatiacatalog.com) -- which includes abundances from individual literature sources that observed stars within 150pc. The Hypatia Catalog currently contains 72 elements as measured within ~6000 stars, with a total of ~240,000 unique abundance determinations. The online database offers a variety of solar normalizations, stellar properties, and planetary properties (where applicable) that can all be viewed through multiple interactive plotting interfaces as well as in a tabular format. By analyzing stellar abundances for large populations of stars and from a variety of different perspectives, a wealth of information can be revealed on both large and small scales.

  13. Software listing: CHEMTOX database

    International Nuclear Information System (INIS)

    Moskowitz, P.D.

    1993-01-01

    Initially launched in 1983, the CHEMTOX Database was among the first microcomputer databases containing hazardous chemical information. The database is used in many industries and government agencies in more than 17 countries. Updated quarterly, the CHEMTOX Database provides detailed environmental and safety information on 7500-plus hazardous substances covered by dozens of regulatory and advisory sources. This brief listing describes the method of accessing data and provides ordering information for those wishing to obtain the CHEMTOX Database

  14. ''Anomalous'' air showers from point sources: Mass limits and light curves

    International Nuclear Information System (INIS)

    Domokos, G.; Elliott, B.; Kovesi-Domokos, S.

    1993-01-01

    We describe a method to obtain upper limits on the mass of the primaries of air showers associated with point sources. One also obtains the UHE pulse shape of a pulsar if its period is observed in the signal. As an example, we analyze the data obtained during a recent burst of Hercules-X1

  15. Coordinating Mobile Databases: A System Demonstration

    OpenAIRE

    Zaihrayeu, Ilya; Giunchiglia, Fausto

    2004-01-01

    In this paper we present the Peer Database Management System (PDBMS). This system runs on top of the standard database management system, and it allows it to connect its database with other (peer) databases on the network. A particularity of our solution is that PDBMS allows for conventional database technology to be effectively operational in mobile settings. We think of database mobility as a database network, where databases appear and disappear spontaneously and their network access point...

  16. Scopus database: a review.

    Science.gov (United States)

    Burnham, Judy F

    2006-03-08

    The Scopus database provides access to STM journal articles and the references included in those articles, allowing the searcher to search both forward and backward in time. The database can be used for collection development as well as for research. This review provides information on the key points of the database and compares it to Web of Science. Neither database is inclusive, but complements each other. If a library can only afford one, choice must be based in institutional needs.

  17. The Atacama Cosmology Telescope: Development and preliminary results of point source observations

    Science.gov (United States)

    Fisher, Ryan P.

    2009-06-01

    The Atacama Cosmology Telescope (ACT) is a six meter diameter telescope designed to measure the millimeter sky with arcminute angular resolution. The instrument is currently conducting its third season of observations from Cerro Toco in the Chilean Andes. The primary science goal of the experiment is to expand our understanding of cosmology by mapping the temperature fluctuations of the Cosmic Microwave Background (CMB) at angular scales corresponding to multipoles up to [cursive l] ~ 10000. The primary receiver for current ACT observations is the Millimeter Bolometer Array Camera (MBAC). The instrument is specially designed to observe simultaneously at 148 GHz, 218 GHz and 277 GHz. To accomplish this, the camera has three separate detector arrays, each containing approximately 1000 detectors. After discussing the ACT experiment in detail, a discussion of the development and testing of the cold readout electronics for the MBAC is presented. Currently, the ACT collaboration is in the process of generating maps of the microwave sky using our first and second season observations. The analysis used to generate these maps requires careful data calibration to produce maps of the arcminute scale CMB temperature fluctuations. Tests and applications of several elements of the ACT calibrations are presented in the context of the second season observations. Scientific exploration has already begun on preliminary maps made using these calibrations. The final portion of this thesis is dedicated to discussing the point sources observed by the ACT. A discussion of the techniques used for point source detection and photometry is followed by a presentation of our current measurements of point source spectral indices.

  18. A 24 μm point source catalog of the galactic plane from Spitzer/MIPSGAL

    Energy Technology Data Exchange (ETDEWEB)

    Gutermuth, Robert A.; Heyer, Mark [Department of Astronomy, University of Massachusetts, Amherst, MA 01003 (United States)

    2015-02-01

    In this contribution, we describe the applied methods to construct a 24 μm based point source catalog derived from the image data of the MIPSGAL 24 μm Galactic Plane Survey and the corresponding data products. The high quality catalog product contains 933,818 sources, with a total of 1,353,228 in the full archive catalog. The source tables include positional and photometric information derived from the 24 μm images, source quality and confusion flags, and counterpart photometry from matched 2MASS, GLIMPSE, and WISE point sources. Completeness decay data cubes are constructed at 1′ angular resolution that describe the varying background levels over the MIPSGAL field and the ability to extract sources of a given magnitude from this background. The completeness decay cubes are included in the set of data products. We present the results of our efforts to verify the astrometric and photometric calibration of the catalog, and present several analyses of minor anomalies in these measurements to justify adopted mitigation strategies.

  19. Source attribution using FLEXPART and carbon monoxide emission inventories for the IAGOS In-situ Observation database

    Science.gov (United States)

    Fontaine, Alain; Sauvage, Bastien; Pétetin, Hervé; Auby, Antoine; Boulanger, Damien; Thouret, Valerie

    2016-04-01

    Since 1994, the IAGOS program (In-Service Aircraft for a Global Observing System http://www.iagos.org) and its predecessor MOZAIC has produced in-situ measurements of the atmospheric composition during more than 46000 commercial aircraft flights. In order to help analyzing these observations and further understanding the processes driving their evolution, we developed a modelling tool SOFT-IO quantifying their source/receptor link. We improved the methodology used by Stohl et al. (2003), based on the FLEXPART plume dispersion model, to simulate the contributions of anthropogenic and biomass burning emissions from the ECCAD database (http://eccad.aeris-data.fr) to the measured carbon monoxide mixing ratio along each IAGOS flight. Thanks to automated processes, contributions are simulated for the last 20 days before observation, separating individual contributions from the different source regions. The main goal is to supply add-value products to the IAGOS database showing pollutants geographical origin and emission type. Using this information, it may be possible to link trends in the atmospheric composition to changes in the transport pathways and to the evolution of emissions. This tool could be used for statistical validation as well as for inter-comparisons of emission inventories using large amounts of data, as Lagrangian models are able to bring the global scale emissions down to a smaller scale, where they can be directly compared to the in-situ observations from the IAGOS database.

  20. LEAP: Looking beyond pixels with continuous-space EstimAtion of Point sources

    Science.gov (United States)

    Pan, Hanjie; Simeoni, Matthieu; Hurley, Paul; Blu, Thierry; Vetterli, Martin

    2017-12-01

    Context. Two main classes of imaging algorithms have emerged in radio interferometry: the CLEAN algorithm and its multiple variants, and compressed-sensing inspired methods. They are both discrete in nature, and estimate source locations and intensities on a regular grid. For the traditional CLEAN-based imaging pipeline, the resolution power of the tool is limited by the width of the synthesized beam, which is inversely proportional to the largest baseline. The finite rate of innovation (FRI) framework is a robust method to find the locations of point-sources in a continuum without grid imposition. The continuous formulation makes the FRI recovery performance only dependent on the number of measurements and the number of sources in the sky. FRI can theoretically find sources below the perceived tool resolution. To date, FRI had never been tested in the extreme conditions inherent to radio astronomy: weak signal / high noise, huge data sets, large numbers of sources. Aims: The aims were (i) to adapt FRI to radio astronomy, (ii) verify it can recover sources in radio astronomy conditions with more accurate positioning than CLEAN, and possibly resolve some sources that would otherwise be missed, (iii) show that sources can be found using less data than would otherwise be required to find them, and (iv) show that FRI does not lead to an augmented rate of false positives. Methods: We implemented a continuous domain sparse reconstruction algorithm in Python. The angular resolution performance of the new algorithm was assessed under simulation, and with visibility measurements from the LOFAR telescope. Existing catalogs were used to confirm the existence of sources. Results: We adapted the FRI framework to radio interferometry, and showed that it is possible to determine accurate off-grid point-source locations and their corresponding intensities. In addition, FRI-based sparse reconstruction required less integration time and smaller baselines to reach a comparable

  1. The Chandra Source Catalog : Automated Source Correlation

    Science.gov (United States)

    Hain, Roger; Evans, I. N.; Evans, J. D.; Glotfelty, K. J.; Anderson, C. S.; Bonaventura, N. R.; Chen, J. C.; Davis, J. E.; Doe, S. M.; Fabbiano, G.; Galle, E.; Gibbs, D. G.; Grier, J. D.; Hall, D. M.; Harbo, P. N.; He, X.; Houck, J. C.; Karovska, M.; Lauer, J.; McCollough, M. L.; McDowell, J. C.; Miller, J. B.; Mitschang, A. W.; Morgan, D. L.; Nichols, J. S.; Nowak, M. A.; Plummer, D. A.; Primini, F. A.; Refsdal, B. L.; Rots, A. H.; Siemiginowska, A. L.; Sundheim, B. A.; Tibbetts, M. S.; Van Stone, D. W.; Winkelman, S. L.; Zografou, P.

    2009-01-01

    Chandra Source Catalog (CSC) master source pipeline processing seeks to automatically detect sources and compute their properties. Since Chandra is a pointed mission and not a sky survey, different sky regions are observed for a different number of times at varying orientations, resolutions, and other heterogeneous conditions. While this provides an opportunity to collect data from a potentially large number of observing passes, it also creates challenges in determining the best way to combine different detection results for the most accurate characterization of the detected sources. The CSC master source pipeline correlates data from multiple observations by updating existing cataloged source information with new data from the same sky region as they become available. This process sometimes leads to relatively straightforward conclusions, such as when single sources from two observations are similar in size and position. Other observation results require more logic to combine, such as one observation finding a single, large source and another identifying multiple, smaller sources at the same position. We present examples of different overlapping source detections processed in the current version of the CSC master source pipeline. We explain how they are resolved into entries in the master source database, and examine the challenges of computing source properties for the same source detected multiple times. Future enhancements are also discussed. This work is supported by NASA contract NAS8-03060 (CXC).

  2. Exact analytical solution of time-independent neutron transport equation, and its applications to systems with a point source

    International Nuclear Information System (INIS)

    Mikata, Y.

    2014-01-01

    Highlights: • An exact solution for the one-speed neutron transport equation is obtained. • This solution as well as its derivation are believed to be new. • Neutron flux for a purely absorbing material with a point neutron source off the origin is obtained. • Spherically as well as cylindrically piecewise constant cross sections are studied. • Neutron flux expressions for a point neutron source off the origin are believed to be new. - Abstract: An exact analytical solution of the time-independent monoenergetic neutron transport equation is obtained in this paper. The solution is applied to systems with a point source. Systematic analysis of the solution of the time-independent neutron transport equation, and its applications represent the primary goal of this paper. To the best of the author’s knowledge, certain key results on the scalar neutron flux as well as their derivations are new. As an application of these results, a scalar neutron flux for a purely absorbing medium with a spherically piecewise constant cross section and an isotropic point neutron source off the origin as well as that for a cylindrically piecewise constant cross section with a point neutron source off the origin are obtained. Both of these results are believed to be new

  3. Integrated Tsunami Database: simulation and identification of seismic tsunami sources, 3D visualization and post-disaster assessment on the shore

    Science.gov (United States)

    Krivorot'ko, Olga; Kabanikhin, Sergey; Marinin, Igor; Karas, Adel; Khidasheli, David

    2013-04-01

    One of the most important problems of tsunami investigation is the problem of seismic tsunami source reconstruction. Non-profit organization WAPMERR (http://wapmerr.org) has provided a historical database of alleged tsunami sources around the world that obtained with the help of information about seaquakes. WAPMERR also has a database of observations of the tsunami waves in coastal areas. The main idea of presentation consists of determining of the tsunami source parameters using seismic data and observations of the tsunami waves on the shore, and the expansion and refinement of the database of presupposed tsunami sources for operative and accurate prediction of hazards and assessment of risks and consequences. Also we present 3D visualization of real-time tsunami wave propagation and loss assessment, characterizing the nature of the building stock in cities at risk, and monitoring by satellite images using modern GIS technology ITRIS (Integrated Tsunami Research and Information System) developed by WAPMERR and Informap Ltd. The special scientific plug-in components are embedded in a specially developed GIS-type graphic shell for easy data retrieval, visualization and processing. The most suitable physical models related to simulation of tsunamis are based on shallow water equations. We consider the initial-boundary value problem in Ω := {(x,y) ?R2 : x ?(0,Lx ), y ?(0,Ly ), Lx,Ly > 0} for the well-known linear shallow water equations in the Cartesian coordinate system in terms of the liquid flow components in dimensional form Here ?(x,y,t) defines the free water surface vertical displacement, i.e. amplitude of a tsunami wave, q(x,y) is the initial amplitude of a tsunami wave. The lateral boundary is assumed to be a non-reflecting boundary of the domain, that is, it allows the free passage of the propagating waves. Assume that the free surface oscillation data at points (xm, ym) are given as a measured output data from tsunami records: fm(t) := ? (xm, ym,t), (xm

  4. Report on the present situation of the FY 1998 technical literature database; 1998 nendo gijutsu bunken database nado genjo chosa

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1999-03-01

    To study database which contributes to the future scientific technology information distribution, survey/analysis were conducted of the present status of the service supply side. In the survey on the database trend, the trend of relations between DB producers and distributors was investigated. As a result, there were seen the increase in DB producers, expansion of internet/distribution/service, etc., and there were no changes in the U.S.-centered structure. Further, it was recognized that the DB service in the internet age now faces the time of change as seen in existing producers' response to internet, on-line service of primary information source, creation of new on-line service, etc. By the internet impact, the following are predicted for the future DB service: slump of producers without strong points and gateway type distributors, appearance of new types of DB service, etc. (NEDO)

  5. Point-Source Contributions to the Water Quality of an Urban Stream

    Science.gov (United States)

    Little, S. F. B.; Young, M.; Lowry, C.

    2014-12-01

    Scajaquada Creek, which runs through the heart of the city of Buffalo, is a prime example of the ways in which human intervention and local geomorphology can impact water quality and urban hydrology. Beginning in the 1920's, the Creek has been partially channelized and connected to Buffalo's combined sewer system (CSS). At Forest Lawn Cemetery, where this study takes place, Scajaquada Creek emerges from a 3.5-mile tunnel built to route stream flow under the city. Collocated with the tunnel outlet is a discharge point for Buffalo's CSS, combined sewer outlet (CSO) #53. It is at this point that runoff and sanitary sewage discharge regularly during rain events. Initially, this study endeavored to create a spatial and temporal picture for this portion of the Creek, monitoring such parameters as conductivity, dissolved oxygen, pH, temperature, and turbidity, in addition to measuring Escherichia coli (E. coli) concentrations. As expected, these factors responded directly to seasonality, local geomorphology, and distance from the point source (CSO #53), displaying a overall, linear response. However, the addition of nitrate and phosphate testing to the study revealed an entirely separate signal from that previously observed. Concentrations of these parameters did not respond to location in the same manner as E. coli. Instead of decreasing with distance from the CSO, a distinct periodicity was observed, correlating with a series of outflow pipes lining the stream banks. It is hypothesized that nitrate and phosphate occurring in this stretch of Scajaquada Creek originate not from the CSO, but from fertilizers used to maintain the lawns within the subwatershed. These results provide evidence of the complexity related to water quality issues in urban streams as a result of point- and nonpoint-source hydrologic inputs.

  6. An international point source outbreak of typhoid fever: a European collaborative investigation*

    Science.gov (United States)

    Stanwell-Smith, R. E.; Ward, L. R.

    1986-01-01

    A point source outbreak of Salmonella typhi, degraded Vi-strain 22, affecting 32 British visitors to Kos, Greece, in 1983 was attributed by a case—control study to the consumption of a salad at one hotel. This represents the first major outbreak of typhoid fever in which a salad has been identified as the vehicle. The source of the infection was probably a carrier in the hotel staff. The investigation demonstrates the importance of national surveillance, international cooperation, and epidemiological methods in the investigation and control of major outbreaks of infection. PMID:3488842

  7. Mercury exposure in terrestrial birds far downstream of an historical point source

    Energy Technology Data Exchange (ETDEWEB)

    Jackson, Allyson K., E-mail: allyson.jackson@briloon.org [Biodiversity Research Institute, 19 Flaggy Meadow Road, Gorham, ME 04038 (United States); Institute for Integrative Bird Behavior Studies, Department of Biology, College of William and Mary, PO Box 8795, Williamsburg, VA 23187 (United States); Evers, David C.; Folsom, Sarah B. [Biodiversity Research Institute, 19 Flaggy Meadow Road, Gorham, ME 04038 (United States); Condon, Anne M. [U.S. Fish and Wildlife Service, 6669 Short Lane, Gloucester, VA 23061 (United States); Diener, John; Goodrick, Lizzie F. [Biodiversity Research Institute, 19 Flaggy Meadow Road, Gorham, ME 04038 (United States); McGann, Andrew J. [Institute for Integrative Bird Behavior Studies, Department of Biology, College of William and Mary, PO Box 8795, Williamsburg, VA 23187 (United States); Schmerfeld, John [U.S. Fish and Wildlife Service, 6669 Short Lane, Gloucester, VA 23061 (United States); Cristol, Daniel A. [Institute for Integrative Bird Behavior Studies, Department of Biology, College of William and Mary, PO Box 8795, Williamsburg, VA 23187 (United States)

    2011-12-15

    Mercury (Hg) is a persistent environmental contaminant found in many freshwater and marine ecosystems. Historical Hg contamination in rivers can impact the surrounding terrestrial ecosystem, but there is little known about how far downstream this contamination persists. In 2009, we sampled terrestrial forest songbirds at five floodplain sites up to 137 km downstream of an historical source of Hg along the South and South Fork Shenandoah Rivers (Virginia, USA). We found that blood total Hg concentrations remained elevated over the entire sampling area and there was little evidence of decline with distance. While it is well known that Hg is a pervasive and long-lasting aquatic contaminant, it has only been recently recognized that it also biomagnifies effectively in floodplain forest food webs. This study extends the area of concern for terrestrial habitats near contaminated rivers for more than 100 km downstream from a waterborne Hg point source. - Highlights: > We report blood mercury levels for terrestrial songbirds downstream of contamination. > Blood mercury levels remain elevated above reference for at least 137 km downstream. > Trends vary based on foraging guild and migration strategy. > Mercury affects terrestrial biota farther downstream than previously documented. - Blood mercury levels of forest songbirds remain elevated above reference levels for at least 137 km downstream of historical point source.

  8. X-Ray and Near-Infrared Spectroscopy of Dim X-Ray Point Sources Constituting the Galactic Ridge X-Ray Emission

    Directory of Open Access Journals (Sweden)

    Kumiko Morihana

    2014-12-01

    Full Text Available We present the results of X-ray and Near-Infrared observations of the Galactic Ridge X-ray Emission (GRXE. We extracted 2,002 X-ray point sources in the Chandra Bulge Field (l =0°.113, b = 1°.424 down to ~10-14.8 ergscm-2s-1 in 2-8 keV band with the longest observation (900 ks of the GRXE. Based on X-ray brightness and hardness, we classied the X-ray point sources into three groups: A (hard, B (soft and broad spectrum, and C (soft and peaked spectrum. In order to know populations of the X-ray point sources, we carried out NIR imaging and spectroscopy observation. We identied 11% of X-ray point sources with NIR and extracted NIR spectra for some of them. Based on X-ray and NIR properties, we concluded that non-thermal sources in the group A are mostly active galactic nuclei and the thermal sources are mostly white dwarf binaries such as cataclysmic variables (CVs and Pre-CVs. We concluded that the group B and C sources are X-ray active stars in flare and quiescence, respectively.

  9. Directory of IAEA databases

    International Nuclear Information System (INIS)

    1991-11-01

    The first edition of the Directory of IAEA Databases is intended to describe the computerized information sources available to IAEA staff members. It contains a listing of all databases produced at the IAEA, together with information on their availability

  10. Non point source pollution modelling in the watershed managed by Integrated Conctructed Wetlands: A GIS approach.

    OpenAIRE

    Vyavahare, Nilesh

    2008-01-01

    The non-point source pollution has been recognised as main cause of eutrophication in Ireland (EPA Ireland, 2001). Integrated Constructed Wetland (ICW) is a management practice adopted in Annestown stream watershed, located in the south county of Waterford in Ireland, used to cleanse farmyard runoff. Present study forms the annual pollution budget for the Annestown stream watershed. The amount of pollution from non-point sources flowing into the stream was simulated by using GIS techniques; u...

  11. OPERA-a human performance database under simulated emergencies of nuclear power plants

    International Nuclear Information System (INIS)

    Park, Jinkyun; Jung, Wondea

    2007-01-01

    In complex systems such as the nuclear and chemical industry, the importance of human performance related problems is well recognized. Thus a lot of effort has been spent on this area, and one of the main streams for unraveling human performance related problems is the execution of HRA. Unfortunately a lack of prerequisite information has been pointed out as the most critical problem in conducting HRA. From this necessity, OPERA database that can provide operators' performance data obtained under simulated emergencies has been developed. In this study, typical operators' performance data that are available from OPERA database are briefly explained. After that, in order to ensure the appropriateness of OPERA database, operators' performance data from OPERA database are compared with those of other studies and real events. As a result, it is believed that operators' performance data of OPERA database are fairly comparable to those of other studies and real events. Therefore it is meaningful to expect that OPERA database can be used as a serviceable data source for scrutinizing human performance related problems including HRA

  12. The recovery of a time-dependent point source in a linear transport equation: application to surface water pollution

    International Nuclear Information System (INIS)

    Hamdi, Adel

    2009-01-01

    The aim of this paper is to localize the position of a point source and recover the history of its time-dependent intensity function that is both unknown and constitutes the right-hand side of a 1D linear transport equation. Assuming that the source intensity function vanishes before reaching the final control time, we prove that recording the state with respect to the time at two observation points framing the source region leads to the identification of the source position and the recovery of its intensity function in a unique manner. Note that at least one of the two observation points should be strategic. We establish an identification method that determines quasi-explicitly the source position and transforms the task of recovering its intensity function into solving directly a well-conditioned linear system. Some numerical experiments done on a variant of the water pollution BOD model are presented

  13. [Validation of interaction databases in psychopharmacotherapy].

    Science.gov (United States)

    Hahn, M; Roll, S C

    2018-03-01

    Drug-drug interaction databases are an important tool to increase drug safety in polypharmacy. There are several drug interaction databases available but it is unclear which one shows the best results and therefore increases safety for the user of the databases and the patients. So far, there has been no validation of German drug interaction databases. Validation of German drug interaction databases regarding the number of hits, mechanisms of drug interaction, references, clinical advice, and severity of the interaction. A total of 36 drug interactions which were published in the last 3-5 years were checked in 5 different databases. Besides the number of hits, it was also documented if the mechanism was correct, clinical advice was given, primary literature was cited, and the severity level of the drug-drug interaction was given. All databases showed weaknesses regarding the hit rate of the tested drug interactions, with a maximum of 67.7% hits. The highest score in this validation was achieved by MediQ with 104 out of 180 points. PsiacOnline achieved 83 points, arznei-telegramm® 58, ifap index® 54 and the ABDA-database 49 points. Based on this validation MediQ seems to be the most suitable databank for the field of psychopharmacotherapy. The best results in this comparison were achieved by MediQ but this database also needs improvement with respect to the hit rate so that the users can rely on the results and therefore increase drug therapy safety.

  14. Comparative Evaluation of Pulsewidth Modulation Strategies for Z-Source Neutral-Point-Clamped Inverter

    DEFF Research Database (Denmark)

    Loh, P.C.; Blaabjerg, Frede; Wong, C.P.

    2007-01-01

    modulation (PWM) strategies for controlling the Z-source NPC inverter. While developing the PWM techniques, attention has been devoted to carefully derive them from a common generic basis for improved portability, easier implementation, and most importantly, assisting readers in understanding all concepts......Z-source neutral-point-clamped (NPC) inverter has recently been proposed as an alternative three-level buck-boost power conversion solution with an improved output waveform quality. In principle, the designed Z-source inverter functions by selectively "shooting through" its power sources, coupled...... to the inverter using two unique Z-source impedance networks, to boost the inverter three-level output waveform. Proper modulation of the new inverter would therefore require careful integration of the selective shoot-through process to the basic switching concepts to achieve maximal voltage-boost, minimal...

  15. Existing data sources for clinical epidemiology: Aarhus University Clinical Trial Candidate Database, Denmark.

    Science.gov (United States)

    Nørrelund, Helene; Mazin, Wiktor; Pedersen, Lars

    2014-01-01

    Denmark is facing a reduction in clinical trial activity as the pharmaceutical industry has moved trials to low-cost emerging economies. Competitiveness in industry-sponsored clinical research depends on speed, quality, and cost. Because Denmark is widely recognized as a region that generates high quality data, an enhanced ability to attract future trials could be achieved if speed can be improved by taking advantage of the comprehensive national and regional registries. A "single point-of-entry" system has been established to support collaboration between hospitals and industry. When assisting industry in early-stage feasibility assessments, potential trial participants are identified by use of registries to shorten the clinical trial startup times. The Aarhus University Clinical Trial Candidate Database consists of encrypted data from the Danish National Registry of Patients allowing an immediate estimation of the number of patients with a specific discharge diagnosis in each hospital department or outpatient specialist clinic in the Central Denmark Region. The free access to health care, thorough monitoring of patients who are in contact with the health service, completeness of registration at the hospital level, and ability to link all databases are competitive advantages in an increasingly complex clinical trial environment.

  16. The Small Area Health Statistics Unit: a national facility for investigating health around point sources of environmental pollution in the United Kingdom.

    Science.gov (United States)

    Elliott, P; Westlake, A J; Hills, M; Kleinschmidt, I; Rodrigues, L; McGale, P; Marshall, K; Rose, G

    1992-01-01

    STUDY OBJECTIVE--The Small Area Health Statistics Unit (SAHSU) was established at the London School of Hygiene and Tropical Medicine in response to a recommendation of the enquiry into the increased incidence of childhood leukaemia near Sellafield, the nuclear reprocessing plant in West Cumbria. The aim of this paper was to describe the Unit's methods for the investigation of health around point sources of environmental pollution in the United Kingdom. DESIGN--Routine data currently including deaths and cancer registrations are held in a large national database which uses a post code based retrieval system to locate cases geographically and link them to the underlying census enumeration districts, and hence to their populations at risk. Main outcome measures were comparison of observed/expected ratios (based on national rates) within bands delineated by concentric circles around point sources of environmental pollution located anywhere in Britain. MAIN RESULTS--The system is illustrated by a study of mortality from mesothelioma and asbestosis near the Plymouth naval dockyards during 1981-87. Within a 3 km radius of the docks the mortality rate for mesothelioma was higher than the national rate by a factor of 8.4, and that for asbestosis was higher by a factor of 13.6. CONCLUSIONS--SAHSU is a new national facility which is rapidly able to provide rates of mortality and cancer incidence for arbitrary circles drawn around any point in Britain. The example around Plymouth of mesothelioma and asbestosis demonstrates the ability of the system to detect an unusual excess of disease in a small locality, although in this case the findings are likely to be related to occupational rather than environmental exposure. PMID:1431704

  17. A proton point source produced by laser interaction with cone-top-end target

    International Nuclear Information System (INIS)

    Yu, Jinqing; Jin, Xiaolin; Zhou, Weimin; Zhao, Zongqing; Yan, Yonghong; Li, Bin; Hong, Wei; Gu, Yuqiu

    2012-01-01

    In this paper, we propose a proton point source by the interaction of laser and cone-top-end target and investigate it by two-dimensional particle-in-cell (2D-PIC) simulations as the proton point sources are well known for higher spatial resolution of proton radiography. Our results show that the relativistic electrons are guided to the rear of the cone-top-end target by the electrostatic charge-separation field and self-generated magnetic field along the profile of the target. As a result, the peak magnitude of sheath field at the rear surface of cone-top-end target is higher compared to common cone target. We test this scheme by 2D-PIC simulation and find the result has a diameter of 0.79λ 0 , an average energy of 9.1 MeV and energy spread less than 35%.

  18. Development of a California commercial building benchmarking database

    International Nuclear Information System (INIS)

    Kinney, Satkartar; Piette, Mary Ann

    2002-01-01

    Building energy benchmarking is a useful starting point for commercial building owners and operators to target energy savings opportunities. There are a number of tools and methods for benchmarking energy use. Benchmarking based on regional data can provides more relevant information for California buildings than national tools such as Energy Star. This paper discusses issues related to benchmarking commercial building energy use and the development of Cal-Arch, a building energy benchmarking database for California. Currently Cal-Arch uses existing survey data from California's Commercial End Use Survey (CEUS), a largely underutilized wealth of information collected by California's major utilities. Doe's Commercial Building Energy Consumption Survey (CBECS) is used by a similar tool, Arch, and by a number of other benchmarking tools. Future versions of Arch/Cal-Arch will utilize additional data sources including modeled data and individual buildings to expand the database

  19. Identifying and characterizing major emission point sources as a basis for geospatial distribution of mercury emissions inventories

    Science.gov (United States)

    Steenhuisen, Frits; Wilson, Simon J.

    2015-07-01

    Mercury is a global pollutant that poses threats to ecosystem and human health. Due to its global transport, mercury contamination is found in regions of the Earth that are remote from major emissions areas, including the Polar regions. Global anthropogenic emission inventories identify important sectors and industries responsible for emissions at a national level; however, to be useful for air transport modelling, more precise information on the locations of emission is required. This paper describes the methodology applied, and the results of work that was conducted to assign anthropogenic mercury emissions to point sources as part of geospatial mapping of the 2010 global anthropogenic mercury emissions inventory prepared by AMAP/UNEP. Major point-source emission sectors addressed in this work account for about 850 tonnes of the emissions included in the 2010 inventory. This work allocated more than 90% of these emissions to some 4600 identified point source locations, including significantly more point source locations in Africa, Asia, Australia and South America than had been identified during previous work to geospatially-distribute the 2005 global inventory. The results demonstrate the utility and the limitations of using existing, mainly public domain resources to accomplish this work. Assumptions necessary to make use of selected online resources are discussed, as are artefacts that can arise when these assumptions are applied to assign (national-sector) emissions estimates to point sources in various countries and regions. Notwithstanding the limitations of the available information, the value of this procedure over alternative methods commonly used to geo-spatially distribute emissions, such as use of 'proxy' datasets to represent emissions patterns, is illustrated. Improvements in information that would facilitate greater use of these methods in future work to assign emissions to point-sources are discussed. These include improvements to both national

  20. Role of rural solid waste management in non-point source pollution control of Dianchi Lake catchments, China

    Institute of Scientific and Technical Information of China (English)

    Wenjing LU; Hongtao WANG

    2008-01-01

    In recent years, with control of the main municipal and industrial point pollution sources and implementation of cleaning for some inner pollution sources in the water body, the discharge of point source pollution decreased gradually, while non-point source pollution has become increasingly distressing in Dianchi Lake catchments. As one of the major targets in non-point source pollution control, an integrated solid waste controlling strategy combined with a technological solution and management system was proposed and implemented based on the waste disposal situation and characteristics of rural solid waste in the demonstration area. As the key technoogy in rural solid waste treatment, both centralized plantscale composting and a dispersed farmer-operated waste treating system showed promise in rendering timely benefits in efficiency, large handling capacity, high quality of the end product, as well as good economic return. Problems encountered during multi-substrates co-com-posting such as pathogens, high moisture content, asyn-chronism in the decomposition of different substrates, and low quality of the end product can all be tackled. 92.5% of solid waste was collected in the demonstration area, while the treating and recycling ratio reached 87.9%, which pre-vented 32.2 t nitrogen and 3.9 t phosphorus per year from entering the water body of Dianchi Lake after imple-mentation of the project.

  1. Using Soluble Reactive Phosphorus and Ammonia to Identify Point Source Discharge from Large Livestock Facilities

    Science.gov (United States)

    Borrello, M. C.; Scribner, M.; Chessin, K.

    2013-12-01

    A growing body of research draws attention to the negative environmental impacts on surface water from large livestock facilities. These impacts are mostly in the form of excessive nutrient loading resulting in significantly decreased oxygen levels. Over-application of animal waste on fields as well as direct discharge into surface water from facilities themselves has been identified as the main contributor to the development of hypoxic zones in Lake Erie, Chesapeake Bay and the Gulf of Mexico. Some regulators claim enforcement of water quality laws is problematic because of the nature and pervasiveness of non-point source impacts. Any direct discharge by a facility is a violation of permits governed by the Clean Water Act, unless the facility has special dispensation for discharge. Previous research by the principal author and others has shown runoff and underdrain transport are the main mechanisms by which nutrients enter surface water. This study utilized previous work to determine if the effects of non-point source discharge can be distinguished from direct (point-source) discharge using simple nutrient analysis and dissolved oxygen (DO) parameters. Nutrient and DO parameters were measured from three sites: 1. A stream adjacent to a field receiving manure, upstream of a large livestock facility with a history of direct discharge, 2. The same stream downstream of the facility and 3. A stream in an area relatively unimpacted by large-scale agriculture (control site). Results show that calculating a simple Pearson correlation coefficient (r) of soluble reactive phosphorus (SRP) and ammonia over time as well as temperature and DO, distinguishes non-point source from point source discharge into surface water. The r value for SRP and ammonia for the upstream site was 0.01 while the r value for the downstream site was 0.92. The control site had an r value of 0.20. Likewise, r values were calculated on temperature and DO for each site. High negative correlations

  2. InverPep: A database of invertebrate antimicrobial peptides.

    Science.gov (United States)

    Gómez, Esteban A; Giraldo, Paula; Orduz, Sergio

    2017-03-01

    The aim of this work was to construct InverPep, a database specialised in experimentally validated antimicrobial peptides (AMPs) from invertebrates. AMP data contained in InverPep were manually curated from other databases and the scientific literature. MySQL was integrated with the development platform Laravel; this framework allows to integrate programming in PHP with HTML and was used to design the InverPep web page's interface. InverPep contains 18 separated fields, including InverPep code, phylum and species source, peptide name, sequence, peptide length, secondary structure, molar mass, charge, isoelectric point, hydrophobicity, Boman index, aliphatic index and percentage of hydrophobic amino acids. CALCAMPI, an algorithm to calculate the physicochemical properties of multiple peptides simultaneously, was programmed in PERL language. To date, InverPep contains 702 experimentally validated AMPs from invertebrate species. All of the peptides contain information associated with their source, physicochemical properties, secondary structure, biological activity and links to external literature. Most AMPs in InverPep have a length between 10 and 50 amino acids, a positive charge, a Boman index between 0 and 2 kcal/mol, and 30-50% hydrophobic amino acids. InverPep includes 33 AMPs not reported in other databases. Besides, CALCAMPI and statistical analysis of InverPep data is presented. The InverPep database is available in English and Spanish. InverPep is a useful database to study invertebrate AMPs and its information could be used for the design of new peptides. The user-friendly interface of InverPep and its information can be freely accessed via a web-based browser at http://ciencias.medellin.unal.edu.co/gruposdeinvestigacion/prospeccionydisenobiomoleculas/InverPep/public/home_en. Copyright © 2016 International Society for Chemotherapy of Infection and Cancer. Published by Elsevier Ltd. All rights reserved.

  3. The 100 strongest radio point sources in the field of the Large Magellanic Cloud at 1.4 GHz

    Directory of Open Access Journals (Sweden)

    Payne J.L.

    2009-01-01

    Full Text Available We present the 100 strongest 1.4 GHz point sources from a new mosaic image in the direction of the Large Magellanic Cloud (LMC. The observations making up the mosaic were made using Australia Telescope Compact Array (ATCA over a ten year period and were combined with Parkes single dish data at 1.4 GHz to complete the image for short spacing. An initial list of co-identifications within 1000 at 0.843, 4.8 and 8.6 GHz consisted of 2682 sources. Elimination of extended objects and artifact noise allowed the creation of a refined list containing 1988 point sources. Most of these are presumed to be background objects seen through the LMC; a small portion may represent compact H ii regions, young SNRs and radio planetary nebulae. For the 1988 point sources we find a preliminary average spectral index (α of -0.53 and present a 1.4 GHz image showing source location in the direction of the LMC.

  4. The 100 Strongest Radio Point Sources in the Field of the Large Magellanic Cloud at 1.4 GHz

    Directory of Open Access Journals (Sweden)

    Payne, J. L.

    2009-06-01

    Full Text Available We present the 100 strongest 1.4~GHz point sources from a new mosaicimage in the direction of the Large Magellanic Cloud (LMC. The observationsmaking up the mosaic were made using Australia Telescope Compact Array (ATCAover a ten year period and were combined with Parkes single dish data at 1.4 GHz to complete the image for short spacing. An initial list of co-identifications within 10arcsec at 0.843, 4.8 and 8.6 GHz consisted of 2682 sources. Elimination of extended objects and artifact noise allowed the creation of a refined list containing 1988 point sources. Most of these are presumed to be background objects seen through the LMC; a small portion may represent compact HII regions, young SNRs and radio planetary nebulae. For the 1988 point sources we find a preliminary average spectral index ($alpha$ of -0.53 and present a 1.4 GHz image showing source locationin the direction of the LMC.

  5. Strategies for lidar characterization of particulates from point and area sources

    Science.gov (United States)

    Wojcik, Michael D.; Moore, Kori D.; Martin, Randal S.; Hatfield, Jerry

    2010-10-01

    Use of ground based remote sensing technologies such as scanning lidar systems (light detection and ranging) has gained traction in characterizing ambient aerosols due to some key advantages such as wide area of regard (10 km2), fast response time, high spatial resolution (University, in conjunction with the USDA-ARS, has developed a three-wavelength scanning lidar system called Aglite that has been successfully deployed to characterize particle motion, concentration, and size distribution at both point and diffuse area sources in agricultural and industrial settings. A suite of massbased and size distribution point sensors are used to locally calibrate the lidar. Generating meaningful particle size distribution, mass concentration, and emission rate results based on lidar data is dependent on strategic onsite deployment of these point sensors with successful local meteorological measurements. Deployment strategies learned from field use of this entire measurement system over five years include the characterization of local meteorology and its predictability prior to deployment, the placement of point sensors to prevent contamination and overloading, the positioning of the lidar and beam plane to avoid hard target interferences, and the usefulness of photographic and written observational data.

  6. Validation of novel calibration scheme with traceable point-like (22)Na sources on six types of PET scanners.

    Science.gov (United States)

    Hasegawa, Tomoyuki; Oda, Keiichi; Wada, Yasuhiro; Sasaki, Toshiaki; Sato, Yasushi; Yamada, Takahiro; Matsumoto, Mikio; Murayama, Hideo; Kikuchi, Kei; Miyatake, Hiroki; Abe, Yutaka; Miwa, Kenta; Akimoto, Kenta; Wagatsuma, Kei

    2013-05-01

    To improve the reliability and convenience of the calibration procedure of positron emission tomography (PET) scanners, we have been developing a novel calibration path based on traceable point-like sources. When using (22)Na sources, special care should be taken to avoid the effects of 1.275-MeV γ rays accompanying β (+) decays. The purpose of this study is to validate this new calibration scheme with traceable point-like (22)Na sources on various types of PET scanners. Traceable point-like (22)Na sources with a spherical absorber design that assures uniform angular distribution of the emitted annihilation photons were used. The tested PET scanners included a clinical whole-body PET scanner, four types of clinical PET/CT scanners from different manufacturers, and a small-animal PET scanner. The region of interest (ROI) diameter dependence of ROI values was represented with a fitting function, which was assumed to consist of a recovery part due to spatial resolution and a quadratic background part originating from the scattered γ rays. The observed ROI radius dependence was well represented with the assumed fitting function (R (2) > 0.994). The calibration factors determined using the point-like sources were consistent with those by the standard cross-calibration method within an uncertainty of ±4 %, which was reasonable considering the uncertainty in the standard cross-calibration method. This novel calibration scheme based on the use of traceable (22)Na point-like sources was successfully validated for six types of commercial PET scanners.

  7. Calibrate the aerial surveying instrument by the limited surface source and the single point source that replace the unlimited surface source

    International Nuclear Information System (INIS)

    Lu Cunheng

    1999-01-01

    It is described that the calculating formula and surveying result is found on the basis of the stacking principle of gamma ray and the feature of hexagonal surface source when the limited surface source replaces the unlimited surface source to calibrate the aerial survey instrument on the ground, and that it is found in the light of the exchanged principle of the gamma ray when the single point source replaces the unlimited surface source to calibrate aerial surveying instrument in the air. Meanwhile through the theoretical analysis, the receiving rate of the crystal bottom and side surfaces is calculated when aerial surveying instrument receives gamma ray. The mathematical expression of the gamma ray decaying following height according to the Jinge function regularity is got. According to this regularity, the absorbing coefficient that air absorbs the gamma ray and the detective efficiency coefficient of the crystal is calculated based on the ground and air measuring value of the bottom surface receiving count rate (derived from total receiving count rate of the bottom and side surface). Finally, according to the measuring value, it is proved that imitating the change of total receiving gamma ray exposure rate of the bottom and side surfaces with this regularity in a certain high area is feasible

  8. A scanning point source for quality control of FOV uniformity in GC-PET imaging

    International Nuclear Information System (INIS)

    Bergmann, H.; Minear, G.; Dobrozemsky, G.; Nowotny, R.; Koenig, B.

    2002-01-01

    Aim: PET imaging with coincidence cameras (GC-PET) requires additional quality control procedures to check the function of coincidence circuitry and detector zoning. In particular, the uniformity response over the field of view needs special attention since it is known that coincidence counting mode may suffer from non-uniformity effects not present in single photon mode. Materials and methods: An inexpensive linear scanner with a stepper motor and a digital interface to a PC with software allowing versatile scanning modes was developed. The scanner is used with a source holder containing a Sodium-22 point source. While moving the source along the axis of rotation of the GC-PET system, a tomographic acquisition takes place. The scan covers the full axial field of view of the 2-D or 3-D scatter frame. Depending on the acquisition software, point source scanning takes place continuously while only one projection is acquired or is done in step-and-shoot mode with the number of positions equal to the number of gantry steps. Special software was developed to analyse the resulting list mode acquisition files and to produce an image of the recorded coincidence events of each head. Results: Uniformity images of coincidence events were obtained after further correction for systematic sensitivity variations caused by acquisition geometry. The resulting images are analysed visually and by calculating NEMA uniformity indices as for a planar flood field. The method has been applied successfully to two different brands of GC-PET capable gamma cameras. Conclusion: Uniformity of GC-PET can be tested quickly and accurately with a routine QC procedure, using a Sodium-22 scanning point source and an inexpensive mechanical scanning device. The method can be used for both 2-D and 3-D acquisition modes and fills an important gap in the quality control system for GC-PET

  9. Mercury exposure in terrestrial birds far downstream of an historical point source

    International Nuclear Information System (INIS)

    Jackson, Allyson K.; Evers, David C.; Folsom, Sarah B.; Condon, Anne M.; Diener, John; Goodrick, Lizzie F.; McGann, Andrew J.; Schmerfeld, John; Cristol, Daniel A.

    2011-01-01

    Mercury (Hg) is a persistent environmental contaminant found in many freshwater and marine ecosystems. Historical Hg contamination in rivers can impact the surrounding terrestrial ecosystem, but there is little known about how far downstream this contamination persists. In 2009, we sampled terrestrial forest songbirds at five floodplain sites up to 137 km downstream of an historical source of Hg along the South and South Fork Shenandoah Rivers (Virginia, USA). We found that blood total Hg concentrations remained elevated over the entire sampling area and there was little evidence of decline with distance. While it is well known that Hg is a pervasive and long-lasting aquatic contaminant, it has only been recently recognized that it also biomagnifies effectively in floodplain forest food webs. This study extends the area of concern for terrestrial habitats near contaminated rivers for more than 100 km downstream from a waterborne Hg point source. - Highlights: → We report blood mercury levels for terrestrial songbirds downstream of contamination. → Blood mercury levels remain elevated above reference for at least 137 km downstream. → Trends vary based on foraging guild and migration strategy. → Mercury affects terrestrial biota farther downstream than previously documented. - Blood mercury levels of forest songbirds remain elevated above reference levels for at least 137 km downstream of historical point source.

  10. Report on the present situation of the FY 1998 technical literature database; 1998 nendo gijutsu bunken database nado genjo chosa

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1999-03-01

    To study database which contributes to the future scientific technology information distribution, survey/analysis were conducted of the present status of the service supply side. In the survey on the database trend, the trend of relations between DB producers and distributors was investigated. As a result, there were seen the increase in DB producers, expansion of internet/distribution/service, etc., and there were no changes in the U.S.-centered structure. Further, it was recognized that the DB service in the internet age now faces the time of change as seen in existing producers' response to internet, on-line service of primary information source, creation of new on-line service, etc. By the internet impact, the following are predicted for the future DB service: slump of producers without strong points and gateway type distributors, appearance of new types of DB service, etc. (NEDO)

  11. A method to analyze “source–sink” structure of non-point source pollution based on remote sensing technology

    International Nuclear Information System (INIS)

    Jiang, Mengzhen; Chen, Haiying; Chen, Qinghui

    2013-01-01

    With the purpose of providing scientific basis for environmental planning about non-point source pollution prevention and control, and improving the pollution regulating efficiency, this paper established the Grid Landscape Contrast Index based on Location-weighted Landscape Contrast Index according to the “source–sink” theory. The spatial distribution of non-point source pollution caused by Jiulongjiang Estuary could be worked out by utilizing high resolution remote sensing images. The results showed that, the area of “source” of nitrogen and phosphorus in Jiulongjiang Estuary was 534.42 km 2 in 2008, and the “sink” was 172.06 km 2 . The “source” of non-point source pollution was distributed mainly over Xiamen island, most of Haicang, east of Jiaomei and river bank of Gangwei and Shima; and the “sink” was distributed over southwest of Xiamen island and west of Shima. Generally speaking, the intensity of “source” gets weaker along with the distance from the seas boundary increase, while “sink” gets stronger. -- Highlights: •We built an index to study the “source–sink” structure of NSP in a space scale. •The Index was applied in Jiulongjiang estuary and got a well result. •The study is beneficial to discern the high load area of non-point source pollution. -- “Source–Sink” Structure of non-point source nitrogen and phosphorus pollution in Jiulongjiang estuary in China was worked out by the Grid Landscape Contrast Index

  12. International patent analysis of water source heat pump based on orbit database

    Science.gov (United States)

    Li, Na

    2018-02-01

    Using orbit database, this paper analysed the international patents of water source heat pump (WSHP) industry with patent analysis methods such as analysis of publication tendency, geographical distribution, technology leaders and top assignees. It is found that the beginning of the 21st century is a period of rapid growth of the patent application of WSHP. Germany and the United States had done researches and development of WSHP in an early time, but now Japan and China have become important countries of patent applications. China has been developing faster and faster in recent years, but the patents are concentrated in universities and urgent to be transferred. Through an objective analysis, this paper aims to provide appropriate decision references for the development of domestic WSHP industry.

  13. Crowd-sourced BMS point matching and metadata maintenance with Babel

    DEFF Research Database (Denmark)

    Fürst, Jonathan; Chen, Kaifei; Katz, Randy H.

    2016-01-01

    Cyber-physical applications, deployed on top of Building Management Systems (BMS), promise energy saving and comfort improvement in non-residential buildings. Such applications are so far mainly deployed as research prototypes. The main roadblock to widespread adoption is the low quality of BMS...... systems. Such applications access sensors and actuators through BMS metadata in form of point labels. The naming of labels is however often inconsistent and incomplete. To tackle this problem, we introduce Babel, a crowd-sourced approach to the creation and maintenance of BMS metadata. In our system...

  14. Effects of pointing compared with naming and observing during encoding on item and source memory in young and older adults.

    Science.gov (United States)

    Ouwehand, Kim; van Gog, Tamara; Paas, Fred

    2016-10-01

    Research showed that source memory functioning declines with ageing. Evidence suggests that encoding visual stimuli with manual pointing in addition to visual observation can have a positive effect on spatial memory compared with visual observation only. The present study investigated whether pointing at picture locations during encoding would lead to better spatial source memory than naming (Experiment 1) and visual observation only (Experiment 2) in young and older adults. Experiment 3 investigated whether response modality during the test phase would influence spatial source memory performance. Experiments 1 and 2 supported the hypothesis that pointing during encoding led to better source memory for picture locations than naming or observation only. Young adults outperformed older adults on the source memory but not the item memory task in both Experiments 1 and 2. In Experiments 1 and 2, participants manually responded in the test phase. Experiment 3 showed that if participants had to verbally respond in the test phase, the positive effect of pointing compared with naming during encoding disappeared. The results suggest that pointing at picture locations during encoding can enhance spatial source memory in both young and older adults, but only if the response modality is congruent in the test phase.

  15. A Comparative Analysis of Vibrio cholerae Contamination in Point-of-Drinking and Source Water in a Low-Income Urban Community, Bangladesh.

    Science.gov (United States)

    Ferdous, Jannatul; Sultana, Rebeca; Rashid, Ridwan B; Tasnimuzzaman, Md; Nordland, Andreas; Begum, Anowara; Jensen, Peter K M

    2018-01-01

    Bangladesh is a cholera endemic country with a population at high risk of cholera. Toxigenic and non-toxigenic Vibrio cholerae ( V. cholerae ) can cause cholera and cholera-like diarrheal illness and outbreaks. Drinking water is one of the primary routes of cholera transmission in Bangladesh. The aim of this study was to conduct a comparative assessment of the presence of V. cholerae between point-of-drinking water and source water, and to investigate the variability of virulence profile using molecular methods of a densely populated low-income settlement of Dhaka, Bangladesh. Water samples were collected and tested for V. cholerae from "point-of-drinking" and "source" in 477 study households in routine visits at 6 week intervals over a period of 14 months. We studied the virulence profiles of V. cholerae positive water samples using 22 different virulence gene markers present in toxigenic O1/O139 and non-O1/O139 V. cholerae using polymerase chain reaction (PCR). A total of 1,463 water samples were collected, with 1,082 samples from point-of-drinking water in 388 households and 381 samples from 66 water sources. V. cholerae was detected in 10% of point-of-drinking water samples and in 9% of source water samples. Twenty-three percent of households and 38% of the sources were positive for V. cholerae in at least one visit. Samples collected from point-of-drinking and linked sources in a 7 day interval showed significantly higher odds ( P source [OR = 17.24 (95% CI = 7.14-42.89)] water. Based on the 7 day interval data, 53% (17/32) of source water samples were negative for V. cholerae while linked point-of-drinking water samples were positive. There were significantly higher odds ( p source water samples than in point-of-drinking water samples. Contamination of water at the point-of-drinking is less likely to depend on the contamination at the water source. Hygiene education interventions and programs should focus and emphasize on water at the point

  16. Existing data sources for clinical epidemiology: the Danish Patient Compensation Association database.

    Science.gov (United States)

    Tilma, Jens; Nørgaard, Mette; Mikkelsen, Kim Lyngby; Johnsen, Søren Paaske

    2015-01-01

    Any patient in the Danish health care system who experiences a treatment injury can make a compensation claim to the Danish Patient Compensation Association (DPCA) free of charge. The aim of this paper is to describe the DPCA database as a source of data for epidemiological research. Data to DPCA are collected prospectively on all claims and include information on patient factors and health records, system factors, and administrative data. Approval of claims is based on injury due to the principle of treatment below experienced specialist standard or intolerable, unexpected extensiveness of injury. Average processing time of a compensation claim is 6-8 months. Data collection is nationwide and started in 1992. The patient's central registration system number, a unique personal identifier, allows for data linkage to other registries such as the Danish National Patient Registry. The DPCA data are accessible for research following data usage permission and make it possible to analyze all claims or specific subgroups to identify predictors, outcomes, etc. DPCA data have until now been used only in few studies but could be a useful data source in future studies of health care-related injuries.

  17. DATABASE REPLICATION IN HETEROGENOUS PLATFORM

    OpenAIRE

    Hendro Nindito; Evaristus Didik Madyatmadja; Albert Verasius Dian Sano

    2014-01-01

    The application of diverse database technologies in enterprises today is increasingly a common practice. To provide high availability and survavibality of real-time information, a database replication technology that has capability to replicate databases under heterogenous platforms is required. The purpose of this research is to find the technology with such capability. In this research, the data source is stored in MSSQL database server running on Windows. The data will be replicated to MyS...

  18. Power-Law Template for IR Point Source Clustering

    Science.gov (United States)

    Addison, Graeme E.; Dunkley, Joanna; Hajian, Amir; Viero, Marco; Bond, J. Richard; Das, Sudeep; Devlin, Mark; Halpern, Mark; Hincks, Adam; Hlozek, Renee; hide

    2011-01-01

    We perform a combined fit to angular power spectra of unresolved infrared (IR) point sources from the Planck satellite (at 217,353,545 and 857 GHz, over angular scales 100 clustered power over the range of angular scales and frequencies considered is well fit by a simple power law of the form C_l\\propto I(sup -n) with n = 1.25 +/- 0.06. While the IR sources are understood to lie at a range of redshifts, with a variety of dust properties, we find that the frequency dependence of the clustering power can be described by the square of a modified blackbody, nu(sup beta) B(nu,T_eff), with a single emissivity index beta = 2.20 +/- 0.07 and effective temperature T_eff= 9.7 K. Our predictions for the clustering amplitude are consistent with existing ACT and South Pole Telescope results at around 150 and 220 GHz, as is our prediction for the effective dust spectral index, which we find to be alpha_150-220 = 3.68 +/- 0.07 between 150 and 220 GHz. Our constraints on the clustering shape and frequency dependence can be used to model the IR clustering as a contaminant in Cosmic Microwave Background anisotropy measurements. The combined Planck and BLAST data also rule out a linear bias clustering model.

  19. Power-Law Template for Infrared Point-Source Clustering

    Science.gov (United States)

    Addison, Graeme E; Dunkley, Joanna; Hajian, Amir; Viero, Marco; Bond, J. Richard; Das, Sudeep; Devlin, Mark J.; Halpern, Mark; Hincks, Adam D; Hlozek, Renee; hide

    2012-01-01

    We perform a combined fit to angular power spectra of unresolved infrared (IR) point sources from the Planck satellite (at 217, 353, 545, and 857 GHz, over angular scales 100 approx clustered power over the range of angular scales and frequencies considered is well fitted by a simple power law of the form C(sup clust)(sub l) varies as l (sub -n) with n = 1.25 +/- 0.06. While the IR sources are understood to lie at a range of redshifts, with a variety of dust properties, we find that the frequency dependence of the clustering power can be described by the square of a modified blackbody, ?(sup Beta)B(?, T(sub eff) ), with a single emissivity index Beta = 2.20 +/- 0.07 and effective temperature T(sub eff) = 9.7 K. Our predictions for the clustering amplitude are consistent with existing ACT and South Pole Telescope results at around 150 and 220 GHz, as is our prediction for the effective dust spectral index, which we find to be alpha(sub 150-220) = 3.68 +/- 0.07 between 150 and 220 GHz. Our constraints on the clustering shape and frequency dependence can be used to model the IR clustering as a contaminant in cosmic microwave background anisotropy measurements. The combined Planck and BLAST data also rule out a linear bias clustering model.

  20. Development of a California commercial building benchmarking database

    Energy Technology Data Exchange (ETDEWEB)

    Kinney, Satkartar; Piette, Mary Ann

    2002-05-17

    Building energy benchmarking is a useful starting point for commercial building owners and operators to target energy savings opportunities. There are a number of tools and methods for benchmarking energy use. Benchmarking based on regional data can provides more relevant information for California buildings than national tools such as Energy Star. This paper discusses issues related to benchmarking commercial building energy use and the development of Cal-Arch, a building energy benchmarking database for California. Currently Cal-Arch uses existing survey data from California's Commercial End Use Survey (CEUS), a largely underutilized wealth of information collected by California's major utilities. Doe's Commercial Building Energy Consumption Survey (CBECS) is used by a similar tool, Arch, and by a number of other benchmarking tools. Future versions of Arch/Cal-Arch will utilize additional data sources including modeled data and individual buildings to expand the database.

  1. [Multiple time scales analysis of spatial differentiation characteristics of non-point source nitrogen loss within watershed].

    Science.gov (United States)

    Liu, Mei-bing; Chen, Xing-wei; Chen, Ying

    2015-07-01

    Identification of the critical source areas of non-point source pollution is an important means to control the non-point source pollution within the watershed. In order to further reveal the impact of multiple time scales on the spatial differentiation characteristics of non-point source nitrogen loss, a SWAT model of Shanmei Reservoir watershed was developed. Based on the simulation of total nitrogen (TN) loss intensity of all 38 subbasins, spatial distribution characteristics of nitrogen loss and critical source areas were analyzed at three time scales of yearly average, monthly average and rainstorms flood process, respectively. Furthermore, multiple linear correlation analysis was conducted to analyze the contribution of natural environment and anthropogenic disturbance on nitrogen loss. The results showed that there were significant spatial differences of TN loss in Shanmei Reservoir watershed at different time scales, and the spatial differentiation degree of nitrogen loss was in the order of monthly average > yearly average > rainstorms flood process. TN loss load mainly came from upland Taoxi subbasin, which was identified as the critical source area. At different time scales, land use types (such as farmland and forest) were always the dominant factor affecting the spatial distribution of nitrogen loss, while the effect of precipitation and runoff on the nitrogen loss was only taken in no fertilization month and several processes of storm flood at no fertilization date. This was mainly due to the significant spatial variation of land use and fertilization, as well as the low spatial variability of precipitation and runoff.

  2. An open, interoperable, transdisciplinary approach to a point cloud data service using OGC standards and open source software.

    Science.gov (United States)

    Steer, Adam; Trenham, Claire; Druken, Kelsey; Evans, Benjamin; Wyborn, Lesley

    2017-04-01

    High resolution point clouds and other topology-free point data sources are widely utilised for research, management and planning activities. A key goal for research and management users is making these data and common derivatives available in a way which is seamlessly interoperable with other observed and modelled data. The Australian National Computational Infrastructure (NCI) stores point data from a range of disciplines, including terrestrial and airborne LiDAR surveys, 3D photogrammetry, airborne and ground-based geophysical observations, bathymetric observations and 4D marine tracers. These data are stored alongside a significant store of Earth systems data including climate and weather, ecology, hydrology, geoscience and satellite observations, and available from NCI's National Environmental Research Data Interoperability Platform (NERDIP) [1]. Because of the NERDIP requirement for interoperability with gridded datasets, the data models required to store these data may not conform to the LAS/LAZ format - the widely accepted community standard for point data storage and transfer. The goal for NCI is making point data discoverable, accessible and useable in ways which allow seamless integration with earth observation datasets and model outputs - in turn assisting researchers and decision-makers in the often-convoluted process of handling and analyzing massive point datasets. With a use-case of providing a web data service and supporting a derived product workflow, NCI has implemented and tested a web-based point cloud service using the Open Geospatial Consortium (OGC) Web Processing Service [2] as a transaction handler between a web-based client and server-side computing tools based on a native Linux operating system. Using this model, the underlying toolset for driving a data service is flexible and can take advantage of NCI's highly scalable research cloud. Present work focusses on the Point Data Abstraction Library (PDAL) [3] as a logical choice for

  3. DIII-D edge physics database

    International Nuclear Information System (INIS)

    Jong, R.A.; Porter, G.D.; Hill, D.N.; Buchenauer, D.A.; Bramson, G.

    1992-03-01

    We have developed an edge-physics database containing data for the plasma in the divertor region and the scrape-off layer (SOL) for the DIII-D tokamak. The database provides many of the parameters necessary to model the power flow to the divertor and other plasma processes in the plasma edge. It will also facilitate the analysis of DIII-D data for comparison with other divertor tokamaks. In addition to the core plasma parameters, edge-specific data are included in this database. Initial results using the database show good agreement between the pressure profiles measured by the Langmuir probes and those determined from the Thomson data for the inner strike point, but not for the outer strike point region. We also find that the ratio of separatrix density to average core density, as well as the in/out asymmetry in the SOL power at the divertor in DIII-D do not agree with values currently assumed in modeling the International Thermonuclear Experimental Reactor (ITER)

  4. Performance of an open-source heart sound segmentation algorithm on eight independent databases.

    Science.gov (United States)

    Liu, Chengyu; Springer, David; Clifford, Gari D

    2017-08-01

    Heart sound segmentation is a prerequisite step for the automatic analysis of heart sound signals, facilitating the subsequent identification and classification of pathological events. Recently, hidden Markov model-based algorithms have received increased interest due to their robustness in processing noisy recordings. In this study we aim to evaluate the performance of the recently published logistic regression based hidden semi-Markov model (HSMM) heart sound segmentation method, by using a wider variety of independently acquired data of varying quality. Firstly, we constructed a systematic evaluation scheme based on a new collection of heart sound databases, which we assembled for the PhysioNet/CinC Challenge 2016. This collection includes a total of more than 120 000 s of heart sounds recorded from 1297 subjects (including both healthy subjects and cardiovascular patients) and comprises eight independent heart sound databases sourced from multiple independent research groups around the world. Then, the HSMM-based segmentation method was evaluated using the assembled eight databases. The common evaluation metrics of sensitivity, specificity, accuracy, as well as the [Formula: see text] measure were used. In addition, the effect of varying the tolerance window for determining a correct segmentation was evaluated. The results confirm the high accuracy of the HSMM-based algorithm on a separate test dataset comprised of 102 306 heart sounds. An average [Formula: see text] score of 98.5% for segmenting S1 and systole intervals and 97.2% for segmenting S2 and diastole intervals were observed. The [Formula: see text] score was shown to increases with an increases in the tolerance window size, as expected. The high segmentation accuracy of the HSMM-based algorithm on a large database confirmed the algorithm's effectiveness. The described evaluation framework, combined with the largest collection of open access heart sound data, provides essential resources for

  5. Temporal-spatial distribution of non-point source pollution in a drinking water source reservoir watershed based on SWAT

    Directory of Open Access Journals (Sweden)

    M. Wang

    2015-05-01

    Full Text Available The conservation of drinking water source reservoirs has a close relationship between regional economic development and people’s livelihood. Research on the non-point pollution characteristics in its watershed is crucial for reservoir security. Tang Pu Reservoir watershed was selected as the study area. The non-point pollution model of Tang Pu Reservoir was established based on the SWAT (Soil and Water Assessment Tool model. The model was adjusted to analyse the temporal-spatial distribution patterns of total nitrogen (TN and total phosphorus (TP. The results showed that the loss of TN and TP in the reservoir watershed were related to precipitation in flood season. And the annual changes showed an "M" shape. It was found that the contribution of loss of TN and TP accounted for 84.5% and 85.3% in high flow years, and for 70.3% and 69.7% in low flow years, respectively. The contributions in normal flow years were 62.9% and 63.3%, respectively. The TN and TP mainly arise from Wangtan town, Gulai town, and Wangyuan town, etc. In addition, it was found that the source of TN and TP showed consistency in space.

  6. Phynx: an open source software solution supporting data management and web-based patient-level data review for drug safety studies in the general practice research database and other health care databases.

    Science.gov (United States)

    Egbring, Marco; Kullak-Ublick, Gerd A; Russmann, Stefan

    2010-01-01

    To develop a software solution that supports management and clinical review of patient data from electronic medical records databases or claims databases for pharmacoepidemiological drug safety studies. We used open source software to build a data management system and an internet application with a Flex client on a Java application server with a MySQL database backend. The application is hosted on Amazon Elastic Compute Cloud. This solution named Phynx supports data management, Web-based display of electronic patient information, and interactive review of patient-level information in the individual clinical context. This system was applied to a dataset from the UK General Practice Research Database (GPRD). Our solution can be setup and customized with limited programming resources, and there is almost no extra cost for software. Access times are short, the displayed information is structured in chronological order and visually attractive, and selected information such as drug exposure can be blinded. External experts can review patient profiles and save evaluations and comments via a common Web browser. Phynx provides a flexible and economical solution for patient-level review of electronic medical information from databases considering the individual clinical context. It can therefore make an important contribution to an efficient validation of outcome assessment in drug safety database studies.

  7. Meta-analysis on Methane Mitigating Properties of Saponin-rich Sources in the Rumen: Influence of Addition Levels and Plant Sources

    Directory of Open Access Journals (Sweden)

    Anuraga Jayanegara

    2014-10-01

    Full Text Available Saponins have been considered as promising natural substances for mitigating methane emissions from ruminants. However, studies reported that addition of saponin-rich sources often arrived at contrasting results, i.e. either it decreased methane or it did not. The aim of the present study was to assess ruminal methane emissions through a meta-analytical approach of integrating related studies from published papers which described various levels of different saponin-rich sources being added to ruminant feed. A database was constructed from published literature reporting the addition of saponin-rich sources at various levels and then monitoring ruminal methane emissions in vitro. Accordingly, levels of saponin-rich source additions as well as different saponin sources were specified in the database. Apart from methane, other related rumen fermentation parameters were also included in the database, i.e. organic matter digestibility, gas production, pH, ammonia concentration, short-chain fatty acid profiles and protozoal count. A total of 23 studies comprised of 89 data points met the inclusion criteria. The data obtained were subsequently subjected to a statistical meta-analysis based on mixed model methodology. Accordingly, different studies were treated as random effects whereas levels of saponin-rich source additions or different saponin sources were considered as fixed effects. Model statistics used were p-value and root mean square error. Results showed that an addition of increasing levels of a saponin-rich source decreased methane emission per unit of substrate incubated as well as per unit of total gas produced (ptea>quillaja, statistically they did not differ each other. It can be concluded that methane mitigating properties of saponins in the rumen are level- and source-dependent.

  8. Airports and Airfields - Volusia County Airports (Points)

    Data.gov (United States)

    NSGIC Local Govt | GIS Inventory — The Public Use Airports database is a geographic point database of aircraft landing facilities in the United States and U.S. Territories. This database has been...

  9. High angle grain boundaries as sources or sinks for point defects

    Energy Technology Data Exchange (ETDEWEB)

    Balluffi, R.W.

    1979-09-01

    A secondary grain boundary dislocation climb model for high angle grain boundaries as sources/sinks for point defects is described in the light of recent advances in our knowledge of grain boundary structure. Experimental results are reviewed and are then compared with the expected behavior of the proposed model. Reasonably good consistency is found at the level of our present understanding of the subject. However, several gaps in our present knowledge still exist, and these are identified and discussed briefly.

  10. The Herschel Virgo Cluster Survey. XVII. SPIRE point-source catalogs and number counts

    Science.gov (United States)

    Pappalardo, Ciro; Bendo, George J.; Bianchi, Simone; Hunt, Leslie; Zibetti, Stefano; Corbelli, Edvige; di Serego Alighieri, Sperello; Grossi, Marco; Davies, Jonathan; Baes, Maarten; De Looze, Ilse; Fritz, Jacopo; Pohlen, Michael; Smith, Matthew W. L.; Verstappen, Joris; Boquien, Médéric; Boselli, Alessandro; Cortese, Luca; Hughes, Thomas; Viaene, Sebastien; Bizzocchi, Luca; Clemens, Marcel

    2015-01-01

    Aims: We present three independent catalogs of point-sources extracted from SPIRE images at 250, 350, and 500 μm, acquired with the Herschel Space Observatory as a part of the Herschel Virgo Cluster Survey (HeViCS). The catalogs have been cross-correlated to consistently extract the photometry at SPIRE wavelengths for each object. Methods: Sources have been detected using an iterative loop. The source positions are determined by estimating the likelihood to be a real source for each peak on the maps, according to the criterion defined in the sourceExtractorSussextractor task. The flux densities are estimated using the sourceExtractorTimeline, a timeline-based point source fitter that also determines the fitting procedure with the width of the Gaussian that best reproduces the source considered. Afterwards, each source is subtracted from the maps, removing a Gaussian function in every position with the full width half maximum equal to that estimated in sourceExtractorTimeline. This procedure improves the robustness of our algorithm in terms of source identification. We calculate the completeness and the flux accuracy by injecting artificial sources in the timeline and estimate the reliability of the catalog using a permutation method. Results: The HeViCS catalogs contain about 52 000, 42 200, and 18 700 sources selected at 250, 350, and 500 μm above 3σ and are ~75%, 62%, and 50% complete at flux densities of 20 mJy at 250, 350, 500 μm, respectively. We then measured source number counts at 250, 350, and 500 μm and compare them with previous data and semi-analytical models. We also cross-correlated the catalogs with the Sloan Digital Sky Survey to investigate the redshift distribution of the nearby sources. From this cross-correlation, we select ~2000 sources with reliable fluxes and a high signal-to-noise ratio, finding an average redshift z ~ 0.3 ± 0.22 and 0.25 (16-84 percentile). Conclusions: The number counts at 250, 350, and 500 μm show an increase in

  11. JICST Factual Database(2)

    Science.gov (United States)

    Araki, Keisuke

    The computer programme, which builds atom-bond connection tables from nomenclatures, is developed. Chemical substances with their nomenclature and varieties of trivial names or experimental code numbers are inputted. The chemical structures of the database are stereospecifically stored and are able to be searched and displayed according to stereochemistry. Source data are from laws and regulations of Japan, RTECS of US and so on. The database plays a central role within the integrated fact database service of JICST and makes interrelational retrieval possible.

  12. Directory of IAEA databases

    International Nuclear Information System (INIS)

    1992-12-01

    This second edition of the Directory of IAEA Databases has been prepared within the Division of Scientific and Technical Information (NESI). Its main objective is to describe the computerized information sources available to staff members. This directory contains all databases produced at the IAEA, including databases stored on the mainframe, LAN's and PC's. All IAEA Division Directors have been requested to register the existence of their databases with NESI. For the second edition database owners were requested to review the existing entries for their databases and answer four additional questions. The four additional questions concerned the type of database (e.g. Bibliographic, Text, Statistical etc.), the category of database (e.g. Administrative, Nuclear Data etc.), the available documentation and the type of media used for distribution. In the individual entries on the following pages the answers to the first two questions (type and category) is always listed, but the answers to the second two questions (documentation and media) is only listed when information has been made available

  13. Zebrafish Database: Customizable, Free, and Open-Source Solution for Facility Management.

    Science.gov (United States)

    Yakulov, Toma Antonov; Walz, Gerd

    2015-12-01

    Zebrafish Database is a web-based customizable database solution, which can be easily adapted to serve both single laboratories and facilities housing thousands of zebrafish lines. The database allows the users to keep track of details regarding the various genomic features, zebrafish lines, zebrafish batches, and their respective locations. Advanced search and reporting options are available. Unique features are the ability to upload files and images that are associated with the respective records and an integrated calendar component that supports multiple calendars and categories. Built on the basis of the Joomla content management system, the Zebrafish Database is easily extendable without the need for advanced programming skills.

  14. Magnox fuel inventories. Experiment and calculation using a point source model

    International Nuclear Information System (INIS)

    Nair, S.

    1978-08-01

    The results of calculations of Magnox fuel inventories using the point source code RICE and associated Magnox reactor data set have been compared with experimental measurements for the actinide isotopes 234 , 235 , 236 , 238 U, 238 , 239 , 240 , 241 , 242 Pu, 241 , 243 Am and 242 , 244 Cm and the fission product isotopes 142 , 143 , 144 , 145 , 146 , 150 Nd, 95 Zr, 134 , 137 Cs, 144 Ce and daughter 144 Pr produced in four samples of spent Magnox fuel spanning the burnup range 3000 to 9000 MWd/Te. The neutron emissions from a further two samples were also measured and compared with RICE predictions. The results of the comparison were such as to justify the use of the code RICE for providing source terms for environmental impact studies, for the isotopes considered in the present work. (author)

  15. Using the Chandra Source-Finding Algorithm to Automatically Identify Solar X-ray Bright Points

    Science.gov (United States)

    Adams, Mitzi L.; Tennant, A.; Cirtain, J. M.

    2009-01-01

    This poster details a technique of bright point identification that is used to find sources in Chandra X-ray data. The algorithm, part of a program called LEXTRCT, searches for regions of a given size that are above a minimum signal to noise ratio. The algorithm allows selected pixels to be excluded from the source-finding, thus allowing exclusion of saturated pixels (from flares and/or active regions). For Chandra data the noise is determined by photon counting statistics, whereas solar telescopes typically integrate a flux. Thus the calculated signal-to-noise ratio is incorrect, but we find we can scale the number to get reasonable results. For example, Nakakubo and Hara (1998) find 297 bright points in a September 11, 1996 Yohkoh image; with judicious selection of signal-to-noise ratio, our algorithm finds 300 sources. To further assess the efficacy of the algorithm, we analyze a SOHO/EIT image (195 Angstroms) and compare results with those published in the literature (McIntosh and Gurman, 2005). Finally, we analyze three sets of data from Hinode, representing different parts of the decline to minimum of the solar cycle.

  16. Databases of the marine metagenomics

    KAUST Repository

    Mineta, Katsuhiko

    2015-10-28

    The metagenomic data obtained from marine environments is significantly useful for understanding marine microbial communities. In comparison with the conventional amplicon-based approach of metagenomics, the recent shotgun sequencing-based approach has become a powerful tool that provides an efficient way of grasping a diversity of the entire microbial community at a sampling point in the sea. However, this approach accelerates accumulation of the metagenome data as well as increase of data complexity. Moreover, when metagenomic approach is used for monitoring a time change of marine environments at multiple locations of the seawater, accumulation of metagenomics data will become tremendous with an enormous speed. Because this kind of situation has started becoming of reality at many marine research institutions and stations all over the world, it looks obvious that the data management and analysis will be confronted by the so-called Big Data issues such as how the database can be constructed in an efficient way and how useful knowledge should be extracted from a vast amount of the data. In this review, we summarize the outline of all the major databases of marine metagenome that are currently publically available, noting that database exclusively on marine metagenome is none but the number of metagenome databases including marine metagenome data are six, unexpectedly still small. We also extend our explanation to the databases, as reference database we call, that will be useful for constructing a marine metagenome database as well as complementing important information with the database. Then, we would point out a number of challenges to be conquered in constructing the marine metagenome database.

  17. KM3NeT/ARCA sensitivity and discovery potential for neutrino point-like sources

    Directory of Open Access Journals (Sweden)

    Trovato A.

    2016-01-01

    Full Text Available KM3NeT is a large research infrastructure with a network of deep-sea neutrino telescopes in the abyss of the Mediterranean Sea. Of these, the KM3NeT/ARCA detector, installed in the KM3NeT-It node of the network, is optimised for studying high-energy neutrinos of cosmic origin. Sensitivities to galactic sources such as the supernova remnant RXJ1713.7-3946 and the pulsar wind nebula Vela X are presented as well as sensitivities to a generic point source with an E−2 spectrum which represents an approximation for the spectrum of extragalactic candidate neutrino sources.

  18. General Approach to the Evolution of Singlet Nanoparticles from a Rapidly Quenched Point Source

    NARCIS (Netherlands)

    Feng, J.; Huang, Luyi; Ludvigsson, Linus; Messing, Maria; Maiser, A.; Biskos, G.; Schmidt-Ott, A.

    2016-01-01

    Among the numerous point vapor sources, microsecond-pulsed spark ablation at atmospheric pressure is a versatile and environmentally friendly method for producing ultrapure inorganic nanoparticles ranging from singlets having sizes smaller than 1 nm to larger agglomerated structures. Due to its fast

  19. Search for neutrino point sources with an all-sky autocorrelation analysis in IceCube

    Energy Technology Data Exchange (ETDEWEB)

    Turcati, Andrea; Bernhard, Anna; Coenders, Stefan [TU, Munich (Germany); Collaboration: IceCube-Collaboration

    2016-07-01

    The IceCube Neutrino Observatory is a cubic kilometre scale neutrino telescope located in the Antarctic ice. Its full-sky field of view gives unique opportunities to study the neutrino emission from the Galactic and extragalactic sky. Recently, IceCube found the first signal of astrophysical neutrinos with energies up to the PeV scale, but the origin of these particles still remains unresolved. Given the observed flux, the absence of observations of bright point-sources is explainable with the presence of numerous weak sources. This scenario can be tested using autocorrelation methods. We present here the sensitivities and discovery potentials of a two-point angular correlation analysis performed on seven years of IceCube data, taken between 2008 and 2015. The test is applied on the northern and southern skies separately, using the neutrino energy information to improve the effectiveness of the method.

  20. Current status of agricultural and rural non-point source Pollution assessment in China

    International Nuclear Information System (INIS)

    Ongley, Edwin D.; Zhang Xiaolan; Yu Tao

    2010-01-01

    Estimates of non-point source (NPS) contribution to total water pollution in China range up to 81% for nitrogen and to 93% for phosphorus. We believe these values are too high, reflecting (a) misuse of estimation techniques that were developed in America under very different conditions and (b) lack of specificity on what is included as NPS. We compare primary methods used for NPS estimation in China with their use in America. Two observations are especially notable: empirical research is limited and does not provide an adequate basis for calibrating models nor for deriving export coefficients; the Chinese agricultural situation is so different than that of the United States that empirical data produced in America, as a basis for applying estimation techniques to rural NPS in China, often do not apply. We propose a set of national research and policy initiatives for future NPS research in China. - Estimation techniques used in China for non-point source pollution are evaluated as a basis for recommending future policies and research in NPS studies in China.

  1. Data-based diffraction kernels for surface waves from convolution and correlation processes through active seismic interferometry

    Science.gov (United States)

    Chmiel, Malgorzata; Roux, Philippe; Herrmann, Philippe; Rondeleux, Baptiste; Wathelet, Marc

    2018-05-01

    We investigated the construction of diffraction kernels for surface waves using two-point convolution and/or correlation from land active seismic data recorded in the context of exploration geophysics. The high density of controlled sources and receivers, combined with the application of the reciprocity principle, allows us to retrieve two-dimensional phase-oscillation diffraction kernels (DKs) of surface waves between any two source or receiver points in the medium at each frequency (up to 15 Hz, at least). These DKs are purely data-based as no model calculations and no synthetic data are needed. They naturally emerge from the interference patterns of the recorded wavefields projected on the dense array of sources and/or receivers. The DKs are used to obtain multi-mode dispersion relations of Rayleigh waves, from which near-surface shear velocity can be extracted. Using convolution versus correlation with a grid of active sources is an important step in understanding the physics of the retrieval of surface wave Green's functions. This provides the foundation for future studies based on noise sources or active sources with a sparse spatial distribution.

  2. Source apportionment of nitrogen and phosphorus from non-point source pollution in Nansi Lake Basin, China.

    Science.gov (United States)

    Zhang, Bao-Lei; Cui, Bo-Hao; Zhang, Shu-Min; Wu, Quan-Yuan; Yao, Lei

    2018-05-03

    Nitrogen (N) and phosphorus (P) from non-point source (NPS) pollution in Nansi Lake Basin greatly influenced the water quality of Nansi Lake, which is the determinant factor for the success of East Route of South-North Water Transfer Project in China. This research improved Johnes export coefficient model (ECM) by developing a method to determine the export coefficients of different land use types based on the hydrological and water quality data. Taking NPS total nitrogen (TN) and total phosphorus (TP) as the study objects, this study estimated the contributions of different pollution sources and analyzed their spatial distributions based on the improved ECM. The results underlined that the method for obtaining output coefficients of land use types using hydrology and water quality data is feasible and accurate, and is suitable for the study of NPS pollution at large-scale basins. The average output structure of NPS TN from land use, rural breeding and rural life is 33.6, 25.9, and 40.5%, and the NPS TP is 31.6, 43.7, and 24.7%, respectively. Especially, dry land was the main land use source for both NPS TN and TP pollution, with the contributed proportions of 81.3 and 81.8% respectively. The counties of Zaozhuang, Tengzhou, Caoxian, Yuncheng, and Shanxian had higher contribution rates and the counties of Dingtao, Juancheng, and Caoxian had the higher load intensities for both NPS TN and TP pollution. The results of this study allowed for an improvement in the understanding of the pollution source contribution and enabled researchers and planners to focus on the most important sources and regions of NPS pollution.

  3. Numerical databases in marine biology

    Digital Repository Service at National Institute of Oceanography (India)

    Sarupria, J.S.; Bhargava, R.M.S.

    stream_size 9 stream_content_type text/plain stream_name Natl_Workshop_Database_Networking_Mar_Biol_1991_45.pdf.txt stream_source_info Natl_Workshop_Database_Networking_Mar_Biol_1991_45.pdf.txt Content-Encoding ISO-8859-1 Content-Type... text/plain; charset=ISO-8859-1 ...

  4. Freshwater Biological Traits Database (Traits)

    Science.gov (United States)

    The traits database was compiled for a project on climate change effects on river and stream ecosystems. The traits data, gathered from multiple sources, focused on information published or otherwise well-documented by trustworthy sources.

  5. A search for hot post-AGE stars in the IRAS Point Source Catalog

    NARCIS (Netherlands)

    Oudmaijer, RD

    In this paper a first step is made to search for hot post-AGB stars in the IRAS Point Source Catalog. In order to find objects that evolved off the AGE a longer time ago than post-AGB objects discussed in the literature, objects that were not detected at 12 mu m by IRAS were selected. The selection

  6. CO2 point sources and subsurface storage capacities for CO2 in aquifers in Norway

    International Nuclear Information System (INIS)

    Boee, Reidulv; Magnus, Christian; Osmundsen, Per Terje; Rindstad, Bjoern Ivar

    2002-01-01

    The GESTCO project comprises a study of the distribution and coincidence of thermal CO 2 emission sources and location/quality of geological storage capacity in Europe. Four of the most promising types of geological storage are being studied. 1. Onshore/offshore saline aquifers with or without lateral seal. 2. Low entalpy geothermal reservoirs. 3. Deep methane-bearing coal beds and abandoned coal and salt mines. 4. Exhausted or near exhausted hydrocarbon structures. In this report we present an inventory of CO 2 point sources in Norway (1999) and the results of the work within Study Area C: Deep saline aquifers offshore/near shore Northern and Central Norway. Also offshore/near shore Southern Norway has been included while the Barents Sea is not described in any detail. The most detailed studies are on the Tilje and Aare Formations on the Troendelag Platform off Mid-Norway and on the Sognefjord, Fensfjord and Krossfjord Formations, southeast of the Troll Field off Western Norway. The Tilje Formation has been chosen as one of the cases to be studied in greater detail (numerical modelling) in the project. This report shows that offshore Norway, there are concentrations of large CO 2 point sources in the Haltenbanken, the Viking Graben/Tampen Spur area, the Southern Viking Graben and the central Trough, while onshore Norway there are concentrations of point sources in the Oslofjord/Porsgrund area, along the coast of western Norway and in the Troendelag. A number of aquifers with large theoretical CO 2 storage potential are pointed out in the North Sea, the Norwegian Sea and in the Southern Barents Sea. The storage capacity in the depth interval 0.8 - 4 km below sea level is estimated to be ca. 13 Gt (13000000000 tonnes) CO 2 in geological traps (outside hydrocarbon fields), while the storage capacity in aquifers not confined to traps is estimated to be at least 280 Gt CO 2 . (Author)

  7. Consumer Product Category Database

    Science.gov (United States)

    The Chemical and Product Categories database (CPCat) catalogs the use of over 40,000 chemicals and their presence in different consumer products. The chemical use information is compiled from multiple sources while product information is gathered from publicly available Material Safety Data Sheets (MSDS). EPA researchers are evaluating the possibility of expanding the database with additional product and use information.

  8. Effect of Uncertainties in CO2 Property Databases on the S-CO2 Compressor Performance

    International Nuclear Information System (INIS)

    Lee, Je Kyoung; Lee, Jeong Ik; Ahn, Yoonhan; Kim, Seong Gu; Cha, Je Eun

    2013-01-01

    Various S-CO 2 Brayton cycle experiment facilities are on the state of construction or operation for demonstration of the technology. However, during the data analysis, S-CO 2 property databases are widely used to predict the performance and characteristics of S-CO 2 Brayton cycle. Thus, a reliable property database is very important before any experiment data analyses or calculation. In this paper, deviation of two different property databases which are widely used for the data analysis will be identified by using three selected properties for comparison, C p , density and enthalpy. Furthermore, effect of above mentioned deviation on the analysis of test data will be briefly discussed. From this deviation, results of the test data analysis can have critical error. As the S-CO 2 Brayton cycle researcher knows, CO 2 near the critical point has dramatic change on thermodynamic properties. Thus, it is true that a potential error source of property prediction exists in CO 2 properties near the critical point. During an experiment data analysis with the S-CO 2 Brayton cycle experiment facility, thermodynamic properties are always involved to predict the component performance and characteristics. Thus, construction or defining of precise CO 2 property database should be carried out to develop Korean S-CO 2 Brayton cycle technology

  9. E-SovTox: An online database of the main publicly-available sources of toxicity data concerning REACH-relevant chemicals published in the Russian language.

    Science.gov (United States)

    Sihtmäe, Mariliis; Blinova, Irina; Aruoja, Villem; Dubourguier, Henri-Charles; Legrand, Nicolas; Kahru, Anne

    2010-08-01

    A new open-access online database, E-SovTox, is presented. E-SovTox provides toxicological data for substances relevant to the EU Registration, Evaluation, Authorisation and Restriction of Chemicals (REACH) system, from publicly-available Russian language data sources. The database contains information selected mainly from scientific journals published during the Soviet Union era. The main information source for this database - the journal, Gigiena Truda i Professional'nye Zabolevania [Industrial Hygiene and Occupational Diseases], published between 1957 and 1992 - features acute, but also chronic, toxicity data for numerous industrial chemicals, e.g. for rats, mice, guinea-pigs and rabbits. The main goal of the abovementioned toxicity studies was to derive the maximum allowable concentration limits for industrial chemicals in the occupational health settings of the former Soviet Union. Thus, articles featured in the database include mostly data on LD50 values, skin and eye irritation, skin sensitisation and cumulative properties. Currently, the E-SovTox database contains toxicity data selected from more than 500 papers covering more than 600 chemicals. The user is provided with the main toxicity information, as well as abstracts of these papers in Russian and in English (given as provided in the original publication). The search engine allows cross-searching of the database by the name or CAS number of the compound, and the author of the paper. The E-SovTox database can be used as a decision-support tool by researchers and regulators for the hazard assessment of chemical substances. 2010 FRAME.

  10. Application of random-point processes to the detection of radiation sources

    International Nuclear Information System (INIS)

    Woods, J.W.

    1978-01-01

    In this report the mathematical theory of random-point processes is reviewed and it is shown how use of the theory can obtain optimal solutions to the problem of detecting radiation sources. As noted, the theory also applies to image processing in low-light-level or low-count-rate situations. Paralleling Snyder's work, the theory is extended to the multichannel case of a continuous, two-dimensional (2-D), energy-time space. This extension essentially involves showing that the data are doubly stochastic Poisson (DSP) point processes in energy as well as time. Further, a new 2-D recursive formulation is presented for the radiation-detection problem with large computational savings over nonrecursive techniques when the number of channels is large (greater than or equal to 30). Finally, some adaptive strategies for on-line ''learning'' of unknown, time-varying signal and background-intensity parameters and statistics are present and discussed. These adaptive procedures apply when a complete statistical description is not available a priori

  11. A New Database Facilitates Characterization of Flavonoid Intake, Sources, and Positive Associations with Diet Quality among US Adults.

    Science.gov (United States)

    Sebastian, Rhonda S; Wilkinson Enns, Cecilia; Goldman, Joseph D; Martin, Carrie L; Steinfeldt, Lois C; Murayi, Theophile; Moshfegh, Alanna J

    2015-06-01

    Epidemiologic studies demonstrate inverse associations between flavonoid intake and chronic disease risk. However, lack of comprehensive databases of the flavonoid content of foods has hindered efforts to fully characterize population intakes and determine associations with diet quality. Using a newly released database of flavonoid values, this study sought to describe intake and sources of total flavonoids and 6 flavonoid classes and identify associations between flavonoid intake and the Healthy Eating Index (HEI) 2010. One day of 24-h dietary recall data from adults aged ≥ 20 y (n = 5420) collected in What We Eat in America (WWEIA), NHANES 2007-2008, were analyzed. Flavonoid intakes were calculated using the USDA Flavonoid Values for Survey Foods and Beverages 2007-2008. Regression analyses were conducted to provide adjusted estimates of flavonoid intake, and linear trends in total and component HEI scores by flavonoid intake were assessed using orthogonal polynomial contrasts. All analyses were weighted to be nationally representative. Mean intake of flavonoids was 251 mg/d, with flavan-3-ols accounting for 81% of intake. Non-Hispanic whites had significantly higher (P empty calories increased (P < 0.001) across flavonoid intake quartiles. A new database that permits comprehensive estimation of flavonoid intakes in WWEIA, NHANES 2007-2008; identification of their major food/beverage sources; and determination of associations with dietary quality will lead to advances in research on relations between flavonoid intake and health. Findings suggest that diet quality, as measured by HEI, is positively associated with flavonoid intake. © 2015 American Society for Nutrition.

  12. Scale out databases for CERN use cases

    International Nuclear Information System (INIS)

    Baranowski, Zbigniew; Grzybek, Maciej; Canali, Luca; Garcia, Daniel Lanza; Surdy, Kacper

    2015-01-01

    Data generation rates are expected to grow very fast for some database workloads going into LHC run 2 and beyond. In particular this is expected for data coming from controls, logging and monitoring systems. Storing, administering and accessing big data sets in a relational database system can quickly become a very hard technical challenge, as the size of the active data set and the number of concurrent users increase. Scale-out database technologies are a rapidly developing set of solutions for deploying and managing very large data warehouses on commodity hardware and with open source software. In this paper we will describe the architecture and tests on database systems based on Hadoop and the Cloudera Impala engine. We will discuss the results of our tests, including tests of data loading and integration with existing data sources and in particular with relational databases. We will report on query performance tests done with various data sets of interest at CERN, notably data from the accelerator log database. (paper)

  13. An Improved Statistical Point-source Foreground Model for the Epoch of Reionization

    Energy Technology Data Exchange (ETDEWEB)

    Murray, S. G.; Trott, C. M.; Jordan, C. H. [ARC Centre of Excellence for All-sky Astrophysics (CAASTRO) (Australia)

    2017-08-10

    We present a sophisticated statistical point-source foreground model for low-frequency radio Epoch of Reionization (EoR) experiments using the 21 cm neutral hydrogen emission line. Motivated by our understanding of the low-frequency radio sky, we enhance the realism of two model components compared with existing models: the source count distributions as a function of flux density and spatial position (source clustering), extending current formalisms for the foreground covariance of 2D power-spectral modes in 21 cm EoR experiments. The former we generalize to an arbitrarily broken power law, and the latter to an arbitrary isotropically correlated field. This paper presents expressions for the modified covariance under these extensions, and shows that for a more realistic source spatial distribution, extra covariance arises in the EoR window that was previously unaccounted for. Failure to include this contribution can yield bias in the final power-spectrum and under-estimate uncertainties, potentially leading to a false detection of signal. The extent of this effect is uncertain, owing to ignorance of physical model parameters, but we show that it is dependent on the relative abundance of faint sources, to the effect that our extension will become more important for future deep surveys. Finally, we show that under some parameter choices, ignoring source clustering can lead to false detections on large scales, due to both the induced bias and an artificial reduction in the estimated measurement uncertainty.

  14. An Improved Statistical Point-source Foreground Model for the Epoch of Reionization

    Science.gov (United States)

    Murray, S. G.; Trott, C. M.; Jordan, C. H.

    2017-08-01

    We present a sophisticated statistical point-source foreground model for low-frequency radio Epoch of Reionization (EoR) experiments using the 21 cm neutral hydrogen emission line. Motivated by our understanding of the low-frequency radio sky, we enhance the realism of two model components compared with existing models: the source count distributions as a function of flux density and spatial position (source clustering), extending current formalisms for the foreground covariance of 2D power-spectral modes in 21 cm EoR experiments. The former we generalize to an arbitrarily broken power law, and the latter to an arbitrary isotropically correlated field. This paper presents expressions for the modified covariance under these extensions, and shows that for a more realistic source spatial distribution, extra covariance arises in the EoR window that was previously unaccounted for. Failure to include this contribution can yield bias in the final power-spectrum and under-estimate uncertainties, potentially leading to a false detection of signal. The extent of this effect is uncertain, owing to ignorance of physical model parameters, but we show that it is dependent on the relative abundance of faint sources, to the effect that our extension will become more important for future deep surveys. Finally, we show that under some parameter choices, ignoring source clustering can lead to false detections on large scales, due to both the induced bias and an artificial reduction in the estimated measurement uncertainty.

  15. Forced sound transmission through a finite-sized single leaf panel subject to a point source excitation.

    Science.gov (United States)

    Wang, Chong

    2018-03-01

    In the case of a point source in front of a panel, the wavefront of the incident wave is spherical. This paper discusses spherical sound waves transmitting through a finite sized panel. The forced sound transmission performance that predominates in the frequency range below the coincidence frequency is the focus. Given the point source located along the centerline of the panel, forced sound transmission coefficient is derived through introducing the sound radiation impedance for spherical incident waves. It is found that in addition to the panel mass, forced sound transmission loss also depends on the distance from the source to the panel as determined by the radiation impedance. Unlike the case of plane incident waves, sound transmission performance of a finite sized panel does not necessarily converge to that of an infinite panel, especially when the source is away from the panel. For practical applications, the normal incidence sound transmission loss expression of plane incident waves can be used if the distance between the source and panel d and the panel surface area S satisfy d/S>0.5. When d/S ≈0.1, the diffuse field sound transmission loss expression may be a good approximation. An empirical expression for d/S=0  is also given.

  16. Jelly Views : Extending Relational Database Systems Toward Deductive Database Systems

    Directory of Open Access Journals (Sweden)

    Igor Wojnicki

    2004-01-01

    Full Text Available This paper regards the Jelly View technology, which provides a new, practical methodology for knowledge decomposition, storage, and retrieval within Relational Database Management Systems (RDBMS. Intensional Knowledge clauses (rules are decomposed and stored in the RDBMS founding reusable components. The results of the rule-based processing are visible as regular views, accessible through SQL. From the end-user point of view the processing capability becomes unlimited (arbitrarily complex queries can be constructed using Intensional Knowledge, while the most external queries are expressed with standard SQL. The RDBMS functionality becomes extended toward that of the Deductive Databases

  17. PRIDE and "Database on Demand" as valuable tools for computational proteomics.

    Science.gov (United States)

    Vizcaíno, Juan Antonio; Reisinger, Florian; Côté, Richard; Martens, Lennart

    2011-01-01

    The Proteomics Identifications Database (PRIDE, http://www.ebi.ac.uk/pride ) provides users with the ability to explore and compare mass spectrometry-based proteomics experiments that reveal details of the protein expression found in a broad range of taxonomic groups, tissues, and disease states. A PRIDE experiment typically includes identifications of proteins, peptides, and protein modifications. Additionally, many of the submitted experiments also include the mass spectra that provide the evidence for these identifications. Finally, one of the strongest advantages of PRIDE in comparison with other proteomics repositories is the amount of metadata it contains, a key point to put the above-mentioned data in biological and/or technical context. Several informatics tools have been developed in support of the PRIDE database. The most recent one is called "Database on Demand" (DoD), which allows custom sequence databases to be built in order to optimize the results from search engines. We describe the use of DoD in this chapter. Additionally, in order to show the potential of PRIDE as a source for data mining, we also explore complex queries using federated BioMart queries to integrate PRIDE data with other resources, such as Ensembl, Reactome, or UniProt.

  18. Assessment of Groundwater Susceptibility to Non-Point Source Contaminants Using Three-Dimensional Transient Indexes.

    Science.gov (United States)

    Zhang, Yong; Weissmann, Gary S; Fogg, Graham E; Lu, Bingqing; Sun, HongGuang; Zheng, Chunmiao

    2018-06-05

    Groundwater susceptibility to non-point source contamination is typically quantified by stable indexes, while groundwater quality evolution (or deterioration globally) can be a long-term process that may last for decades and exhibit strong temporal variations. This study proposes a three-dimensional (3- d ), transient index map built upon physical models to characterize the complete temporal evolution of deep aquifer susceptibility. For illustration purposes, the previous travel time probability density (BTTPD) approach is extended to assess the 3- d deep groundwater susceptibility to non-point source contamination within a sequence stratigraphic framework observed in the Kings River fluvial fan (KRFF) aquifer. The BTTPD, which represents complete age distributions underlying a single groundwater sample in a regional-scale aquifer, is used as a quantitative, transient measure of aquifer susceptibility. The resultant 3- d imaging of susceptibility using the simulated BTTPDs in KRFF reveals the strong influence of regional-scale heterogeneity on susceptibility. The regional-scale incised-valley fill deposits increase the susceptibility of aquifers by enhancing rapid downward solute movement and displaying relatively narrow and young age distributions. In contrast, the regional-scale sequence-boundary paleosols within the open-fan deposits "protect" deep aquifers by slowing downward solute movement and displaying a relatively broad and old age distribution. Further comparison of the simulated susceptibility index maps to known contaminant distributions shows that these maps are generally consistent with the high concentration and quick evolution of 1,2-dibromo-3-chloropropane (DBCP) in groundwater around the incised-valley fill since the 1970s'. This application demonstrates that the BTTPDs can be used as quantitative and transient measures of deep aquifer susceptibility to non-point source contamination.

  19. Non-point Source Pollutants Loss of Planting Industry in the Yunnan Plateau Lake Basin, China

    Directory of Open Access Journals (Sweden)

    ZHAO Zu-jun

    2017-12-01

    Full Text Available Non-point source pollution of planting has become a major factor affecting the quality and safety of water environment in our country. In recent years, some studies show that the loss of nitrogen and phosphorus in agricultural chemical fertilizers has led to more serious non-point source pollution. By means of the loss coefficient method and spatial overlay analysis, the loss amount, loss of strength and its spatial distribution characteristics of total nitrogen, total phosphorus, ammonium nitrogen and nitrate nitrogen were analyzed in the Fuxian Lake, Xingyun Lake and Qilu Lake Basin in 2015. The results showed that:The loss of total nitrogen was the highest in the three basins, following by ammonium nitrogen, nitrate nitrogen and total phosphorus, which the loss of intensity range were 2.73~22.07, 0.003~3.52, 0.01~2.25 kg·hm-2 and 0.05~1.36 kg·hm-2, respectively. Total nitrogen and total phosphorus loss were mainly concentrated in the southwest of Qilu Lake, west and south of Xingyun Lake. Ammonium nitrogen and nitrate nitrogen loss mainly concentrated in the south of Qilu Lake, south and north of Xingyun Lake. The loss of nitrogen and phosphorus was mainly derived from cash crops and rice. Therefore, zoning, grading and phased prevention and control schemes were proposed, in order to provide scientific basis for controlling non-point source pollution in the study area.

  20. TRAM (Transcriptome Mapper: database-driven creation and analysis of transcriptome maps from multiple sources

    Directory of Open Access Journals (Sweden)

    Danieli Gian

    2011-02-01

    Full Text Available Abstract Background Several tools have been developed to perform global gene expression profile data analysis, to search for specific chromosomal regions whose features meet defined criteria as well as to study neighbouring gene expression. However, most of these tools are tailored for a specific use in a particular context (e.g. they are species-specific, or limited to a particular data format and they typically accept only gene lists as input. Results TRAM (Transcriptome Mapper is a new general tool that allows the simple generation and analysis of quantitative transcriptome maps, starting from any source listing gene expression values for a given gene set (e.g. expression microarrays, implemented as a relational database. It includes a parser able to assign univocal and updated gene symbols to gene identifiers from different data sources. Moreover, TRAM is able to perform intra-sample and inter-sample data normalization, including an original variant of quantile normalization (scaled quantile, useful to normalize data from platforms with highly different numbers of investigated genes. When in 'Map' mode, the software generates a quantitative representation of the transcriptome of a sample (or of a pool of samples and identifies if segments of defined lengths are over/under-expressed compared to the desired threshold. When in 'Cluster' mode, the software searches for a set of over/under-expressed consecutive genes. Statistical significance for all results is calculated with respect to genes localized on the same chromosome or to all genome genes. Transcriptome maps, showing differential expression between two sample groups, relative to two different biological conditions, may be easily generated. We present the results of a biological model test, based on a meta-analysis comparison between a sample pool of human CD34+ hematopoietic progenitor cells and a sample pool of megakaryocytic cells. Biologically relevant chromosomal segments and gene

  1. Gamma Rays from the Inner Milky Way: Dark Matter or Point Sources?

    CERN Multimedia

    CERN. Geneva

    2015-01-01

    Studies of data from the Fermi Gamma-Ray Space Telescope have revealed bright gamma-ray emission from the central regions of our galaxy, with a spatial and spectral profile consistent with annihilating dark matter. I will present a new model-independent analysis that suggests that rather than originating from dark matter, the GeV excess may arise from a surprising new population of as-yet-unresolved gamma-ray point sources in the heart of the Milky Way.

  2. The Hanford Site generic component failure-rate database compared with other generic failure-rate databases

    International Nuclear Information System (INIS)

    Reardon, M.F.; Zentner, M.D.

    1992-11-01

    The Risk Assessment Technology Group, Westinghouse Hanford Company (WHC), has compiled a component failure rate database to be used during risk and reliability analysis of nonreactor facilities. Because site-specific data for the Hanford Site are generally not kept or not compiled in a usable form, the database was assembled using information from a variety of other established sources. Generally, the most conservative failure rates were chosen from the databases reviewed. The Hanford Site database has since been used extensively in fault tree modeling of many Hanford Site facilities and systems. The purpose of this study was to evaluate the reasonableness of the data chosen for the Hanford Site database by comparing the values chosen with the values from the other databases

  3. Identification of a time-varying point source in a system of two coupled linear diffusion-advection- reaction equations: application to surface water pollution

    International Nuclear Information System (INIS)

    Hamdi, Adel

    2009-01-01

    This paper deals with the identification of a point source (localization of its position and recovering the history of its time-varying intensity function) that constitutes the right-hand side of the first equation in a system of two coupled 1D linear transport equations. Assuming that the source intensity function vanishes before reaching the final control time, we prove the identifiability of the sought point source from recording the state relative to the second coupled transport equation at two observation points framing the source region. Note that at least one of the two observation points should be strategic. We establish an identification method that uses these records to identify the source position as the root of a continuous and strictly monotonic function. Whereas the source intensity function is recovered using a recursive formula without any need of an iterative process. Some numerical experiments on a variant of the surface water pollution BOD–OD coupled model are presented

  4. High frequency seismic signal generated by landslides on complex topographies: from point source to spatially distributed sources

    Science.gov (United States)

    Mangeney, A.; Kuehnert, J.; Capdeville, Y.; Durand, V.; Stutzmann, E.; Kone, E. H.; Sethi, S.

    2017-12-01

    During their flow along the topography, landslides generate seismic waves in a wide frequency range. These so called landquakes can be recorded at very large distances (a few hundreds of km for large landslides). The recorded signals depend on the landslide seismic source and the seismic wave propagation. If the wave propagation is well understood, the seismic signals can be inverted for the seismic source and thus can be used to get information on the landslide properties and dynamics. Analysis and modeling of long period seismic signals (10-150s) have helped in this way to discriminate between different landslide scenarios and to constrain rheological parameters (e.g. Favreau et al., 2010). This was possible as topography poorly affects wave propagation at these long periods and the landslide seismic source can be approximated as a point source. In the near-field and at higher frequencies (> 1 Hz) the spatial extent of the source has to be taken into account and the influence of the topography on the recorded seismic signal should be quantified in order to extract information on the landslide properties and dynamics. The characteristic signature of distributed sources and varying topographies is studied as a function of frequency and recording distance.The time dependent spatial distribution of the forces applied to the ground by the landslide are obtained using granular flow numerical modeling on 3D topography. The generated seismic waves are simulated using the spectral element method. The simulated seismic signal is compared to observed seismic data from rockfalls at the Dolomieu Crater of Piton de la Fournaise (La Réunion).Favreau, P., Mangeney, A., Lucas, A., Crosta, G., and Bouchut, F. (2010). Numerical modeling of landquakes. Geophysical Research Letters, 37(15):1-5.

  5. Identification of 'Point A' as the prevalent source of error in cephalometric analysis of lateral radiographs.

    Science.gov (United States)

    Grogger, P; Sacher, C; Weber, S; Millesi, G; Seemann, R

    2018-04-10

    Deviations in measuring dentofacial components in a lateral X-ray represent a major hurdle in the subsequent treatment of dysgnathic patients. In a retrospective study, we investigated the most prevalent source of error in the following commonly used cephalometric measurements: the angles Sella-Nasion-Point A (SNA), Sella-Nasion-Point B (SNB) and Point A-Nasion-Point B (ANB); the Wits appraisal; the anteroposterior dysplasia indicator (APDI); and the overbite depth indicator (ODI). Preoperative lateral radiographic images of patients with dentofacial deformities were collected and the landmarks digitally traced by three independent raters. Cephalometric analysis was automatically performed based on 1116 tracings. Error analysis identified the x-coordinate of Point A as the prevalent source of error in all investigated measurements, except SNB, in which it is not incorporated. In SNB, the y-coordinate of Nasion predominated error variance. SNB showed lowest inter-rater variation. In addition, our observations confirmed previous studies showing that landmark identification variance follows characteristic error envelopes in the highest number of tracings analysed up to now. Variance orthogonal to defining planes was of relevance, while variance parallel to planes was not. Taking these findings into account, orthognathic surgeons as well as orthodontists would be able to perform cephalometry more accurately and accomplish better therapeutic results. Copyright © 2018 International Association of Oral and Maxillofacial Surgeons. Published by Elsevier Ltd. All rights reserved.

  6. The IAEA Illicit Trafficking Database Programme: Operations and Structure

    International Nuclear Information System (INIS)

    2010-01-01

    the IAEA I TDB currently has 90 states participating voluntarily to the database. Information on about 827 incidents of which 500 involved radioactive sources has been reported. States provide information by submitting an Information Notification Form. The incident is assigned an identification number and entered in the database. Information from open sources is collected daily and reviewed. If the information warrants it a new incident is created in the database.

  7. International Shock-Wave Database: Current Status

    Science.gov (United States)

    Levashov, Pavel

    2013-06-01

    Shock-wave and related dynamic material response data serve for calibrating, validating, and improving material models over very broad regions of the pressure-temperature-density phase space. Since the middle of the 20th century vast amount of shock-wave experimental information has been obtained. To systemize it a number of compendiums of shock-wave data has been issued by LLNL, LANL (USA), CEA (France), IPCP and VNIIEF (Russia). In mid-90th the drawbacks of the paper handbooks became obvious, so the first version of the online shock-wave database appeared in 1997 (http://www.ficp.ac.ru/rusbank). It includes approximately 20000 experimental points on shock compression, adiabatic expansion, measurements of sound velocity behind the shock front and free-surface-velocity for more than 650 substances. This is still a useful tool for the shock-wave community, but it has a number of serious disadvantages which can't be easily eliminated: (i) very simple data format for points and references; (ii) minimalistic user interface for data addition; (iii) absence of history of changes; (iv) bad feedback from users. The new International Shock-Wave database (ISWdb) is intended to solve these and some other problems. The ISWdb project objectives are: (i) to develop a database on thermodynamic and mechanical properties of materials under conditions of shock-wave and other dynamic loadings, selected related quantities of interest, and the meta-data that describes the provenance of the measurements and material models; and (ii) to make this database available internationally through the Internet, in an interactive form. The development and operation of the ISWdb is guided by an advisory committee. The database will be installed on two mirrored web-servers, one in Russia and the other in USA (currently only one server is available). The database provides access to original experimental data on shock compression, non-shock dynamic loadings, isentropic expansion, measurements of sound

  8. A Targeted Search for Point Sources of EeV Photons with the Pierre Auger Observatory

    Energy Technology Data Exchange (ETDEWEB)

    Aab, A. [Institute for Mathematics, Astrophysics and Particle Physics (IMAPP), Radboud Universiteit, Nijmegen (Netherlands); Abreu, P. [Laboratório de Instrumentação e Física Experimental de Partículas—LIP and Instituto Superior Técnico—IST, Universidade de Lisboa—UL, Lisbon (Portugal); Aglietta, M. [INFN, Sezione di Torino, Torino (Italy); Samarai, I. Al [Laboratoire de Physique Nucléaire et de Hautes Energies (LPNHE), Universités Paris 6 et Paris 7, CNRS-IN2P3, Paris (France); Albuquerque, I. F. M. [Universidade de São Paulo, Inst. de Física, São Paulo (Brazil); Allekotte, I. [Centro Atómico Bariloche and Instituto Balseiro (CNEA-UNCuyo-CONICET), San Carlos de Bariloche (Argentina); Almela, A. [Instituto de Tecnologías en Detección y Astropartículas (CNEA, CONICET, UNSAM), Centro Atómico Constituyentes, Comisión Nacional de Energía Atómica, Buenos Aires (Argentina); Castillo, J. Alvarez [Universidad Nacional Autónoma de México, México, D. F., México (Mexico); Alvarez-Muñiz, J. [Universidad de Santiago de Compostela, La Coruña (Spain); Anastasi, G. A. [Gran Sasso Science Institute (INFN), L’Aquila (Italy); and others

    2017-03-10

    Simultaneous measurements of air showers with the fluorescence and surface detectors of the Pierre Auger Observatory allow a sensitive search for EeV photon point sources. Several Galactic and extragalactic candidate objects are grouped in classes to reduce the statistical penalty of many trials from that of a blind search and are analyzed for a significant excess above the background expectation. The presented search does not find any evidence for photon emission at candidate sources, and combined p -values for every class are reported. Particle and energy flux upper limits are given for selected candidate sources. These limits significantly constrain predictions of EeV proton emission models from non-transient Galactic and nearby extragalactic sources, as illustrated for the particular case of the Galactic center region.

  9. Methane Flux Estimation from Point Sources using GOSAT Target Observation: Detection Limit and Improvements with Next Generation Instruments

    Science.gov (United States)

    Kuze, A.; Suto, H.; Kataoka, F.; Shiomi, K.; Kondo, Y.; Crisp, D.; Butz, A.

    2017-12-01

    Atmospheric methane (CH4) has an important role in global radiative forcing of climate but its emission estimates have larger uncertainties than carbon dioxide (CO2). The area of anthropogenic emission sources is usually much smaller than 100 km2. The Thermal And Near infrared Sensor for carbon Observation Fourier-Transform Spectrometer (TANSO-FTS) onboard the Greenhouse gases Observing SATellite (GOSAT) has measured CO2 and CH4 column density using sun light reflected from the earth's surface. It has an agile pointing system and its footprint can cover 87-km2 with a single detector. By specifying pointing angles and observation time for every orbit, TANSO-FTS can target various CH4 point sources together with reference points every 3 day over years. We selected a reference point that represents CH4 background density before or after targeting a point source. By combining satellite-measured enhancement of the CH4 column density and surface measured wind data or estimates from the Weather Research and Forecasting (WRF) model, we estimated CH4emission amounts. Here, we picked up two sites in the US West Coast, where clear sky frequency is high and a series of data are available. The natural gas leak at Aliso Canyon showed a large enhancement and its decrease with time since the initial blowout. We present time series of flux estimation assuming the source is single point without influx. The observation of the cattle feedlot in Chino, California has weather station within the TANSO-FTS footprint. The wind speed is monitored continuously and the wind direction is stable at the time of GOSAT overpass. The large TANSO-FTS footprint and strong wind decreases enhancement below noise level. Weak wind shows enhancements in CH4, but the velocity data have large uncertainties. We show the detection limit of single samples and how to reduce uncertainty using time series of satellite data. We will propose that the next generation instruments for accurate anthropogenic CO2 and CH

  10. DISCRIMINATION OF NATURAL AND NON-POINT SOURCE EFFECTS FROM ANTHROGENIC EFFECTS AS REFLECTED IN BENTHIC STATE IN THREE ESTUARIES IN NEW ENGLAND

    Science.gov (United States)

    In order to protect estuarine resources, managers must be able to discern the effects of natural conditions and non-point source effects, and separate them from multiple anthropogenic point source effects. Our approach was to evaluate benthic community assemblages, riverine nitro...

  11. UHE γ-rays from point sources based on GRAPES-I observations

    International Nuclear Information System (INIS)

    Gupta, S.K.; Sreekantan, B.V.; Srivatsan, R.; Tonwar, S.C.

    1993-01-01

    An experiment called GRAPES I (Gamma Ray Astronomy at PeV EnergieS) was set up in 1984 at Ooty in India, using 24 scintillation counters, to detect Extensive Air Showers (EAS) produced in the atmosphere by the primary cosmic radiation. The goal of the experiment has been to search for Ultra High Energy (UHE) γ-rays (E≥10 14 eV) from point sources in the sky. Here we discuss the results on X-ray binaries CYG X-3, HER X-1 and SCO X-1 obtained with GRAPES I experiment which covers the period 1984--87

  12. POWER-LAW TEMPLATE FOR INFRARED POINT-SOURCE CLUSTERING

    Energy Technology Data Exchange (ETDEWEB)

    Addison, Graeme E.; Dunkley, Joanna [Sub-department of Astrophysics, University of Oxford, Denys Wilkinson Building, Keble Road, Oxford OX1 3RH (United Kingdom); Hajian, Amir; Das, Sudeep; Hincks, Adam D.; Page, Lyman A.; Staggs, Suzanne T. [Joseph Henry Laboratories of Physics, Jadwin Hall, Princeton University, Princeton, NJ 08544 (United States); Viero, Marco [Department of Astronomy, California Institute of Technology, 1200 East California Boulevard, Pasadena, CA 91125 (United States); Bond, J. Richard [Canadian Institute for Theoretical Astrophysics, University of Toronto, Toronto, ON M5S 3H8 (Canada); Devlin, Mark J.; Reese, Erik D. [Department of Physics and Astronomy, University of Pennsylvania, 209 South 33rd Street, Philadelphia, PA 19104 (United States); Halpern, Mark; Scott, Douglas [Department of Physics and Astronomy, University of British Columbia, Vancouver, BC V6T 1Z4 (Canada); Hlozek, Renee; Marriage, Tobias A.; Spergel, David N. [Department of Astrophysical Sciences, Peyton Hall, Princeton University, Princeton, NJ 08544 (United States); Moodley, Kavilan [Astrophysics and Cosmology Research Unit, School of Mathematical Sciences, University of KwaZulu-Natal, Durban 4041 (South Africa); Wollack, Edward [NASA/Goddard Space Flight Center, Code 665, Greenbelt, MD 20771 (United States)

    2012-06-20

    We perform a combined fit to angular power spectra of unresolved infrared (IR) point sources from the Planck satellite (at 217, 353, 545, and 857 GHz, over angular scales 100 {approx}< l {approx}< 2200), the Balloon-borne Large-Aperture Submillimeter Telescope (BLAST; 250, 350, and 500 {mu}m; 1000 {approx}< l {approx}< 9000), and from correlating BLAST and Atacama Cosmology Telescope (ACT; 148 and 218 GHz) maps. We find that the clustered power over the range of angular scales and frequencies considered is well fitted by a simple power law of the form C{sup clust}{sub l}{proportional_to}l{sup -n} with n = 1.25 {+-} 0.06. While the IR sources are understood to lie at a range of redshifts, with a variety of dust properties, we find that the frequency dependence of the clustering power can be described by the square of a modified blackbody, {nu}{sup {beta}} B({nu}, T{sub eff}), with a single emissivity index {beta} = 2.20 {+-} 0.07 and effective temperature T{sub eff} = 9.7 K. Our predictions for the clustering amplitude are consistent with existing ACT and South Pole Telescope results at around 150 and 220 GHz, as is our prediction for the effective dust spectral index, which we find to be {alpha}{sub 150-220} = 3.68 {+-} 0.07 between 150 and 220 GHz. Our constraints on the clustering shape and frequency dependence can be used to model the IR clustering as a contaminant in cosmic microwave background anisotropy measurements. The combined Planck and BLAST data also rule out a linear bias clustering model.

  13. Atmospheric mercury dispersion modelling from two nearest hypothetical point sources

    Energy Technology Data Exchange (ETDEWEB)

    Al Razi, Khandakar Md Habib; Hiroshi, Moritomi; Shinji, Kambara [Environmental and Renewable Energy System (ERES), Graduate School of Engineering, Gifu University, Yanagido, Gifu City, 501-1193 (Japan)

    2012-07-01

    The Japan coastal areas are still environmentally friendly, though there are multiple air emission sources originating as a consequence of several developmental activities such as automobile industries, operation of thermal power plants, and mobile-source pollution. Mercury is known to be a potential air pollutant in the region apart from SOX, NOX, CO and Ozone. Mercury contamination in water bodies and other ecosystems due to deposition of atmospheric mercury is considered a serious environmental concern. Identification of sources contributing to the high atmospheric mercury levels will be useful for formulating pollution control and mitigation strategies in the region. In Japan, mercury and its compounds were categorized as hazardous air pollutants in 1996 and are on the list of 'Substances Requiring Priority Action' published by the Central Environmental Council of Japan. The Air Quality Management Division of the Environmental Bureau, Ministry of the Environment, Japan, selected the current annual mean environmental air quality standard for mercury and its compounds of 0.04 ?g/m3. Long-term exposure to mercury and its compounds can have a carcinogenic effect, inducing eg, Minamata disease. This study evaluates the impact of mercury emissions on air quality in the coastal area of Japan. Average yearly emission of mercury from an elevated point source in this area with background concentration and one-year meteorological data were used to predict the ground level concentration of mercury. To estimate the concentration of mercury and its compounds in air of the local area, two different simulation models have been used. The first is the National Institute of Advanced Science and Technology Atmospheric Dispersion Model for Exposure and Risk Assessment (AIST-ADMER) that estimates regional atmospheric concentration and distribution. The second is the Hybrid Single Particle Lagrangian Integrated trajectory Model (HYSPLIT) that estimates the atmospheric

  14. Field validation of secondary data sources: a novel measure of representativity applied to a Canadian food outlet database.

    Science.gov (United States)

    Clary, Christelle M; Kestens, Yan

    2013-06-19

    Validation studies of secondary datasets used to characterize neighborhood food businesses generally evaluate how accurately the database represents the true situation on the ground. Depending on the research objectives, the characterization of the business environment may tolerate some inaccuracies (e.g. minor imprecisions in location or errors in business names). Furthermore, if the number of false negatives (FNs) and false positives (FPs) is balanced within a given area, one could argue that the database still provides a "fair" representation of existing resources in this area. Yet, traditional validation measures do not relax matching criteria, and treat FNs and FPs independently. Through the field validation of food businesses found in a Canadian database, this paper proposes alternative criteria for validity. Field validation of the 2010 Enhanced Points of Interest (EPOI) database (DMTI Spatial®) was performed in 2011 in 12 census tracts (CTs) in Montreal, Canada. Some 410 food outlets were extracted from the database and 484 were observed in the field. First, traditional measures of sensitivity and positive predictive value (PPV) accounting for every single mismatch between the field and the database were computed. Second, relaxed measures of sensitivity and PPV that tolerate mismatches in business names or slight imprecisions in location were assessed. A novel measure of representativity that further allows for compensation between FNs and FPs within the same business category and area was proposed. Representativity was computed at CT level as ((TPs +|FPs-FNs|)/(TPs+FNs)), with TPs meaning true positives, and |FPs-FNs| being the absolute value of the difference between the number of FNs and the number of FPs within each outlet category. The EPOI database had a "moderate" capacity to detect an outlet present in the field (sensitivity: 54.5%) or to list only the outlets that actually existed in the field (PPV: 64.4%). Relaxed measures of sensitivity and PPV

  15. Super-resolution for a point source better than λ/500 using positive refraction

    Science.gov (United States)

    Miñano, Juan C.; Marqués, Ricardo; González, Juan C.; Benítez, Pablo; Delgado, Vicente; Grabovickic, Dejan; Freire, Manuel

    2011-12-01

    Leonhardt (2009 New J. Phys. 11 093040) demonstrated that the two-dimensional (2D) Maxwell fish eye (MFE) lens can focus perfectly 2D Helmholtz waves of arbitrary frequency; that is, it can transport perfectly an outward (monopole) 2D Helmholtz wave field, generated by a point source, towards a ‘perfect point drain’ located at the corresponding image point. Moreover, a prototype with λ/5 super-resolution property for one microwave frequency has been manufactured and tested (Ma et al 2010 arXiv:1007.2530v1; Ma et al 2010 New J. Phys. 13 033016). However, neither software simulations nor experimental measurements for a broad band of frequencies have yet been reported. Here, we present steady-state simulations with a non-perfect drain for a device equivalent to the MFE, called the spherical geodesic waveguide (SGW), which predicts up to λ/500 super-resolution close to discrete frequencies. Out of these frequencies, the SGW does not show super-resolution in the analysis carried out.

  16. Super-resolution for a point source better than λ/500 using positive refraction

    International Nuclear Information System (INIS)

    Miñano, Juan C; González, Juan C; Benítez, Pablo; Grabovickic, Dejan; Marqués, Ricardo; Delgado, Vicente; Freire, Manuel

    2011-01-01

    Leonhardt (2009 New J. Phys. 11 093040) demonstrated that the two-dimensional (2D) Maxwell fish eye (MFE) lens can focus perfectly 2D Helmholtz waves of arbitrary frequency; that is, it can transport perfectly an outward (monopole) 2D Helmholtz wave field, generated by a point source, towards a ‘perfect point drain’ located at the corresponding image point. Moreover, a prototype with λ/5 super-resolution property for one microwave frequency has been manufactured and tested (Ma et al 2010 arXiv:1007.2530v1; Ma et al 2010 New J. Phys. 13 033016). However, neither software simulations nor experimental measurements for a broad band of frequencies have yet been reported. Here, we present steady-state simulations with a non-perfect drain for a device equivalent to the MFE, called the spherical geodesic waveguide (SGW), which predicts up to λ/500 super-resolution close to discrete frequencies. Out of these frequencies, the SGW does not show super-resolution in the analysis carried out. (paper)

  17. Programmable current source for diode lasers stabilized optical fiber

    International Nuclear Information System (INIS)

    Gomez, J.; Camas, J.; Garcia, L.

    2012-01-01

    In this paper, we present the electronic design of a programmable stabilized current source. User can access to the source through a password, which, it has a database with the current and voltage operating points. This source was successfully used as current source in laser diode in optical fiber sensors. Variations in the laser current were carried out by a monitoring system and a control of the Direct Current (DC), which flowing through a How land source with amplifier. The laser current can be stabilized with an error percent of ± 1 μA from the threshold current (Ith) to its maximum operation current (Imax) in DC mode. The proposed design is reliable, cheap, and its output signal of stabilized current has high quality. (Author)

  18. Correlation Wave-Front Sensing Algorithms for Shack-Hartmann-Based Adaptive Optics using a Point Source

    International Nuclear Information System (INIS)

    Poynee, L A

    2003-01-01

    Shack-Hartmann based Adaptive Optics system with a point-source reference normally use a wave-front sensing algorithm that estimates the centroid (center of mass) of the point-source image 'spot' to determine the wave-front slope. The centroiding algorithm suffers for several weaknesses. For a small number of pixels, the algorithm gain is dependent on spot size. The use of many pixels on the detector leads to significant propagation of read noise. Finally, background light or spot halo aberrations can skew results. In this paper an alternative algorithm that suffers from none of these problems is proposed: correlation of the spot with a ideal reference spot. The correlation method is derived and a theoretical analysis evaluates its performance in comparison with centroiding. Both simulation and data from real AO systems are used to illustrate the results. The correlation algorithm is more robust than centroiding, but requires more computation

  19. Major and Trace Element Fluxes to the Ganges River: Significance of Small Flood Plain Tributary as Non-Point Pollution Source

    Science.gov (United States)

    Lakshmi, V.; Sen, I. S.; Mishra, G.

    2017-12-01

    There has been much discussion amongst biologists, ecologists, chemists, geologists, environmental firms, and science policy makers about the impact of human activities on river health. As a result, multiple river restoration projects are on going on many large river basins around the world. In the Indian subcontinent, the Ganges River is the focal point of all restoration actions as it provides food and water security to half a billion people. Serious concerns have been raised about the quality of Ganga water as toxic chemicals and many more enters the river system through point-sources such as direct wastewater discharge to rivers, or non-point-sources. Point source pollution can be easily identified and remedial actions can be taken; however, non-point pollution sources are harder to quantify and mitigate. A large non-point pollution source in the Indo-Gangetic floodplain is the network of small floodplain rivers. However, these rivers are rarely studied since they are small in catchment area ( 1000-10,000 km2) and discharge (knowledge gap we have monitored the Pandu River for one year between February 2015 and April 2016. Pandu river is 242 km long and is a right bank tributary of Ganges with a total catchment area of 1495 km2. Water samples were collected every month for dissolved major and trace elements. Here we show that the concentration of heavy metals in river Pandu is in higher range as compared to the world river average, and all the dissolved elements shows a large spatial-temporal variation. We show that the Pandu river exports 192170, 168517, 57802, 32769, 29663, 1043, 279, 241, 225, 162, 97, 28, 25, 22, 20, 8, 4 Kg/yr of Ca, Na, Mg, K, Si, Sr, Zn, B, Ba, Mn, Al, Li, Rb, Mo, U, Cu, and Sb, respectively, to the Ganga river, and the exported chemical flux effects the water chemistry of the Ganga river downstream of its confluence point. We further speculate that small floodplain rivers is an important source that contributes to the dissolved chemical

  20. Super-resolution for a point source using positive refraction

    Science.gov (United States)

    Miñano, Juan C.; Benítez, Pablo; González, Juan C.; Grabovičkić, Dejan; Ahmadpanahi, Hamed

    Leonhardt demonstrated (2009) that the 2D Maxwell Fish Eye lens (MFE) can focus perfectly 2D Helmholtz waves of arbitrary frequency, i.e., it can transport perfectly an outward (monopole) 2D Helmholtz wave field, generated by a point source, towards a receptor called "perfect drain" (PD) located at the corresponding MFE image point. The PD has the property of absorbing the complete radiation without radiation or scattering and it has been claimed as necessary to obtain super-resolution (SR) in the MFE. However, a prototype using a "drain" different from the PD has shown λ/5 resolution for microwave frequencies (Ma et al, 2010). Recently, the SR properties of a device equivalent to the MFE, called the Spherical Geodesic Waveguide (SGW) (Miñano et al, 2012) have been analyzed. The reported results show resolution up to λ /3000, for the SGW loaded with the perfect drain, and up to λ /500 for the SGW without perfect drain. The perfect drain was realized as a coaxial probe loaded with properly calculated impedance. The SGW provides SR only in a narrow band of frequencies close to the resonance Schumann frequencies. Here we analyze the SGW loaded with a small "perfect drain region" (González et al, 2011). This drain is designed as a region made of a material with complex permittivity. The comparative results show that there is no significant difference in the SR properties for both perfect drain designs.

  1. CHANDRA ACIS SURVEY OF X-RAY POINT SOURCES: THE SOURCE CATALOG

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Song; Liu, Jifeng; Qiu, Yanli; Bai, Yu; Yang, Huiqin; Guo, Jincheng; Zhang, Peng, E-mail: jfliu@bao.ac.cn, E-mail: songw@bao.ac.cn [Key Laboratory of Optical Astronomy, National Astronomical Observatories, Chinese Academy of Sciences, Beijing 100012 (China)

    2016-06-01

    The Chandra archival data is a valuable resource for various studies on different X-ray astronomy topics. In this paper, we utilize this wealth of information and present a uniformly processed data set, which can be used to address a wide range of scientific questions. The data analysis procedures are applied to 10,029 Advanced CCD Imaging Spectrometer observations, which produces 363,530 source detections belonging to 217,828 distinct X-ray sources. This number is twice the size of the Chandra Source Catalog (Version 1.1). The catalogs in this paper provide abundant estimates of the detected X-ray source properties, including source positions, counts, colors, fluxes, luminosities, variability statistics, etc. Cross-correlation of these objects with galaxies shows that 17,828 sources are located within the D {sub 25} isophotes of 1110 galaxies, and 7504 sources are located between the D {sub 25} and 2 D {sub 25} isophotes of 910 galaxies. Contamination analysis with the log N –log S relation indicates that 51.3% of objects within 2 D {sub 25} isophotes are truly relevant to galaxies, and the “net” source fraction increases to 58.9%, 67.3%, and 69.1% for sources with luminosities above 10{sup 37}, 10{sup 38}, and 10{sup 39} erg s{sup −1}, respectively. Among the possible scientific uses of this catalog, we discuss the possibility of studying intra-observation variability, inter-observation variability, and supersoft sources (SSSs). About 17,092 detected sources above 10 counts are classified as variable in individual observation with the Kolmogorov–Smirnov (K–S) criterion ( P {sub K–S} < 0.01). There are 99,647 sources observed more than once and 11,843 sources observed 10 times or more, offering us a wealth of data with which to explore the long-term variability. There are 1638 individual objects (∼2350 detections) classified as SSSs. As a quite interesting subclass, detailed studies on X-ray spectra and optical spectroscopic follow-up are needed to

  2. Analysis of large databases in vascular surgery.

    Science.gov (United States)

    Nguyen, Louis L; Barshes, Neal R

    2010-09-01

    Large databases can be a rich source of clinical and administrative information on broad populations. These datasets are characterized by demographic and clinical data for over 1000 patients from multiple institutions. Since they are often collected and funded for other purposes, their use for secondary analysis increases their utility at relatively low costs. Advantages of large databases as a source include the very large numbers of available patients and their related medical information. Disadvantages include lack of detailed clinical information and absence of causal descriptions. Researchers working with large databases should also be mindful of data structure design and inherent limitations to large databases, such as treatment bias and systemic sampling errors. Withstanding these limitations, several important studies have been published in vascular care using large databases. They represent timely, "real-world" analyses of questions that may be too difficult or costly to address using prospective randomized methods. Large databases will be an increasingly important analytical resource as we focus on improving national health care efficacy in the setting of limited resources.

  3. Discretized energy minimization in a wave guide with point sources

    Science.gov (United States)

    Propst, G.

    1994-01-01

    An anti-noise problem on a finite time interval is solved by minimization of a quadratic functional on the Hilbert space of square integrable controls. To this end, the one-dimensional wave equation with point sources and pointwise reflecting boundary conditions is decomposed into a system for the two propagating components of waves. Wellposedness of this system is proved for a class of data that includes piecewise linear initial conditions and piecewise constant forcing functions. It is shown that for such data the optimal piecewise constant control is the solution of a sparse linear system. Methods for its computational treatment are presented as well as examples of their applicability. The convergence of discrete approximations to the general optimization problem is demonstrated by finite element methods.

  4. A scalable and multi-purpose point cloud server (PCS) for easier and faster point cloud data management and processing

    Science.gov (United States)

    Cura, Rémi; Perret, Julien; Paparoditis, Nicolas

    2017-05-01

    In addition to more traditional geographical data such as images (rasters) and vectors, point cloud data are becoming increasingly available. Such data are appreciated for their precision and true three-Dimensional (3D) nature. However, managing point clouds can be difficult due to scaling problems and specificities of this data type. Several methods exist but are usually fairly specialised and solve only one aspect of the management problem. In this work, we propose a comprehensive and efficient point cloud management system based on a database server that works on groups of points (patches) rather than individual points. This system is specifically designed to cover the basic needs of point cloud users: fast loading, compressed storage, powerful patch and point filtering, easy data access and exporting, and integrated processing. Moreover, the proposed system fully integrates metadata (like sensor position) and can conjointly use point clouds with other geospatial data, such as images, vectors, topology and other point clouds. Point cloud (parallel) processing can be done in-base with fast prototyping capabilities. Lastly, the system is built on open source technologies; therefore it can be easily extended and customised. We test the proposed system with several billion points obtained from Lidar (aerial and terrestrial) and stereo-vision. We demonstrate loading speeds in the ˜50 million pts/h per process range, transparent-for-user and greater than 2 to 4:1 compression ratio, patch filtering in the 0.1 to 1 s range, and output in the 0.1 million pts/s per process range, along with classical processing methods, such as object detection.

  5. Selection of nuclear power information database management system

    International Nuclear Information System (INIS)

    Zhang Shuxin; Wu Jianlei

    1996-01-01

    In the condition of the present database technology, in order to build the Chinese nuclear power information database (NPIDB) in the nuclear industry system efficiently at a high starting point, an important task is to select a proper database management system (DBMS), which is the hinge of the matter to build the database successfully. Therefore, this article explains how to build a practical information database about nuclear power, the functions of different database management systems, the reason of selecting relation database management system (RDBMS), the principles of selecting RDBMS, the recommendation of ORACLE management system as the software to build database and so on

  6. Scale out databases for CERN use cases

    CERN Document Server

    Baranowski, Zbigniew; Canali, Luca; Garcia, Daniel Lanza; Surdy, Kacper

    2015-01-01

    Data generation rates are expected to grow very fast for some database workloads going into LHC run 2 and beyond. In particular this is expected for data coming from controls, logging and monitoring systems. Storing, administering and accessing big data sets in a relational database system can quickly become a very hard technical challenge, as the size of the active data set and the number of concurrent users increase. Scale-out database technologies are a rapidly developing set of solutions for deploying and managing very large data warehouses on commodity hardware and with open source software. In this paper we will describe the architecture and tests on database systems based on Hadoop and the Cloudera Impala engine. We will discuss the results of our tests, including tests of data loading and integration with existing data sources and in particular with relational databases. We will report on query performance tests done with various data sets of interest at CERN, notably data from the accelerator log dat...

  7. The Development and Application of Spatiotemporal Metrics for the Characterization of Point Source FFCO2 Emissions and Dispersion

    Science.gov (United States)

    Roten, D.; Hogue, S.; Spell, P.; Marland, E.; Marland, G.

    2017-12-01

    There is an increasing role for high resolution, CO2 emissions inventories across multiple arenas. The breadth of the applicability of high-resolution data is apparent from their use in atmospheric CO2 modeling, their potential for validation of space-based atmospheric CO2 remote-sensing, and the development of climate change policy. This work focuses on increasing our understanding of the uncertainty in these inventories and the implications on their downstream use. The industrial point sources of emissions (power generating stations, cement manufacturing plants, paper mills, etc.) used in the creation of these inventories often have robust emissions characteristics, beyond just their geographic location. Physical parameters of the emission sources such as number of exhaust stacks, stack heights, stack diameters, exhaust temperatures, and exhaust velocities, as well as temporal variability and climatic influences can be important in characterizing emissions. Emissions from large point sources can behave much differently than emissions from areal sources such as automobiles. For many applications geographic location is not an adequate characterization of emissions. This work demonstrates the sensitivities of atmospheric models to the physical parameters of large point sources and provides a methodology for quantifying parameter impacts at multiple locations across the United States. The sensitivities highlight the importance of location and timing and help to highlight potential aspects that can guide efforts to reduce uncertainty in emissions inventories and increase the utility of the models.

  8. Towards P2P XML Database Technology

    NARCIS (Netherlands)

    Y. Zhang (Ying)

    2007-01-01

    textabstractTo ease the development of data-intensive P2P applications, we envision a P2P XML Database Management System (P2P XDBMS) that acts as a database middle-ware, providing a uniform database abstraction on top of a dynamic set of distributed data sources. In this PhD work, we research which

  9. Online Sources of Competitive Intelligence.

    Science.gov (United States)

    Wagers, Robert

    1986-01-01

    Presents an approach to using online sources of information for competitor intelligence (i.e., monitoring industry and tracking activities of competitors); identifies principal sources; and suggests some ways of making use of online databases. Types and sources of information and sources and database charts are appended. Eight references are…

  10. Monte Carlo simulation of scatter in non-uniform symmetrical attenuating media for point and distributed sources

    International Nuclear Information System (INIS)

    Henry, L.J.; Rosenthal, M.S.

    1992-01-01

    We report results of scatter simulations for both point and distributed sources of 99m Tc in symmetrical non-uniform attenuating media. The simulations utilized Monte Carlo techniques and were tested against experimental phantoms. Both point and ring sources were used inside a 10.5 cm radius acrylic phantom. Attenuating media consisted of combinations of water, ground beef (to simulate muscle mass), air and bone meal (to simulate bone mass). We estimated/measured energy spectra, detector efficiencies and peak height ratios for all cases. In all cases, the simulated spectra agree with the experimentally measured spectra within 2 SD. Detector efficiencies and peak height ratios also are in agreement. The Monte Carlo code is able to properly model the non-uniform attenuating media used in this project. With verification of the simulations, it is possible to perform initial evaluation studies of scatter correction algorithms by evaluating the mechanisms of action of the correction algorithm on the simulated spectra where the magnitude and sources of scatter are known. (author)

  11. An Integrated Enterprise Accelerator Database for the SLC Control System

    International Nuclear Information System (INIS)

    2002-01-01

    Since its inception in the early 1980's, the SLC Control System has been driven by a highly structured memory-resident real-time database. While efficient, its rigid structure and file-based sources makes it difficult to maintain and extract relevant information. The goal of transforming the sources for this database into a relational form is to enable it to be part of a Control System Enterprise Database that is an integrated central repository for SLC accelerator device and Control System data with links to other associated databases. We have taken the concepts developed for the NLC Enterprise Database and used them to create and load a relational model of the online SLC Control System database. This database contains data and structure to allow querying and reporting on beamline devices, their associations and parameters. In the future this will be extended to allow generation of EPICS and SLC database files, setup of applications and links to other databases such as accelerator maintenance, archive data, financial and personnel records, cabling information, documentation etc. The database is implemented using Oracle 8i. In the short term it will be updated daily in batch from the online SLC database. In the longer term, it will serve as the primary source for Control System static data, an R and D platform for the NLC, and contribute to SLC Control System operations

  12. Transformer-based asymmetrical embedded Z-source neutral point clamped inverters with continuous input current and enhanced voltage boost capability

    DEFF Research Database (Denmark)

    Mo, W.; Loh, Poh Chiang; Blaabjerg, Frede

    2013-01-01

    Z-source Neutral Point Clamped (NPC) inverters were introduced to integrate both the advantages of Z-source inverters and NPC inverters. However, traditional Z-source inverters suffer from high voltage stress and chopping input current. This paper proposes six types transformer-based impedance-so......-source NPC inverters which have enhanced voltage boost capability and continuous input current by utilizing of transformer and embedded dc source configuration. Experimental results are presented to verify the theory validation....

  13. Building a multi-scaled geospatial temporal ecology database from disparate data sources: Fostering open science through data reuse

    Science.gov (United States)

    Soranno, Patricia A.; Bissell, E.G.; Cheruvelil, Kendra S.; Christel, Samuel T.; Collins, Sarah M.; Fergus, C. Emi; Filstrup, Christopher T.; Lapierre, Jean-Francois; Lotting, Noah R.; Oliver, Samantha K.; Scott, Caren E.; Smith, Nicole J.; Stopyak, Scott; Yuan, Shuai; Bremigan, Mary Tate; Downing, John A.; Gries, Corinna; Henry, Emily N.; Skaff, Nick K.; Stanley, Emily H.; Stow, Craig A.; Tan, Pang-Ning; Wagner, Tyler; Webster, Katherine E.

    2015-01-01

    Although there are considerable site-based data for individual or groups of ecosystems, these datasets are widely scattered, have different data formats and conventions, and often have limited accessibility. At the broader scale, national datasets exist for a large number of geospatial features of land, water, and air that are needed to fully understand variation among these ecosystems. However, such datasets originate from different sources and have different spatial and temporal resolutions. By taking an open-science perspective and by combining site-based ecosystem datasets and national geospatial datasets, science gains the ability to ask important research questions related to grand environmental challenges that operate at broad scales. Documentation of such complicated database integration efforts, through peer-reviewed papers, is recommended to foster reproducibility and future use of the integrated database. Here, we describe the major steps, challenges, and considerations in building an integrated database of lake ecosystems, called LAGOS (LAke multi-scaled GeOSpatial and temporal database), that was developed at the sub-continental study extent of 17 US states (1,800,000 km2). LAGOS includes two modules: LAGOSGEO, with geospatial data on every lake with surface area larger than 4 ha in the study extent (~50,000 lakes), including climate, atmospheric deposition, land use/cover, hydrology, geology, and topography measured across a range of spatial and temporal extents; and LAGOSLIMNO, with lake water quality data compiled from ~100 individual datasets for a subset of lakes in the study extent (~10,000 lakes). Procedures for the integration of datasets included: creating a flexible database design; authoring and integrating metadata; documenting data provenance; quantifying spatial measures of geographic data; quality-controlling integrated and derived data; and extensively documenting the database. Our procedures make a large, complex, and integrated

  14. Building a multi-scaled geospatial temporal ecology database from disparate data sources: fostering open science and data reuse.

    Science.gov (United States)

    Soranno, Patricia A; Bissell, Edward G; Cheruvelil, Kendra S; Christel, Samuel T; Collins, Sarah M; Fergus, C Emi; Filstrup, Christopher T; Lapierre, Jean-Francois; Lottig, Noah R; Oliver, Samantha K; Scott, Caren E; Smith, Nicole J; Stopyak, Scott; Yuan, Shuai; Bremigan, Mary Tate; Downing, John A; Gries, Corinna; Henry, Emily N; Skaff, Nick K; Stanley, Emily H; Stow, Craig A; Tan, Pang-Ning; Wagner, Tyler; Webster, Katherine E

    2015-01-01

    Although there are considerable site-based data for individual or groups of ecosystems, these datasets are widely scattered, have different data formats and conventions, and often have limited accessibility. At the broader scale, national datasets exist for a large number of geospatial features of land, water, and air that are needed to fully understand variation among these ecosystems. However, such datasets originate from different sources and have different spatial and temporal resolutions. By taking an open-science perspective and by combining site-based ecosystem datasets and national geospatial datasets, science gains the ability to ask important research questions related to grand environmental challenges that operate at broad scales. Documentation of such complicated database integration efforts, through peer-reviewed papers, is recommended to foster reproducibility and future use of the integrated database. Here, we describe the major steps, challenges, and considerations in building an integrated database of lake ecosystems, called LAGOS (LAke multi-scaled GeOSpatial and temporal database), that was developed at the sub-continental study extent of 17 US states (1,800,000 km(2)). LAGOS includes two modules: LAGOSGEO, with geospatial data on every lake with surface area larger than 4 ha in the study extent (~50,000 lakes), including climate, atmospheric deposition, land use/cover, hydrology, geology, and topography measured across a range of spatial and temporal extents; and LAGOSLIMNO, with lake water quality data compiled from ~100 individual datasets for a subset of lakes in the study extent (~10,000 lakes). Procedures for the integration of datasets included: creating a flexible database design; authoring and integrating metadata; documenting data provenance; quantifying spatial measures of geographic data; quality-controlling integrated and derived data; and extensively documenting the database. Our procedures make a large, complex, and integrated

  15. [Urban non-point source pollution control by runoff retention and filtration pilot system].

    Science.gov (United States)

    Bai, Yao; Zuo, Jian-E; Gan, Li-Li; Low, Thong Soon; Miao, Heng-Feng; Ruan, Wen-Quan; Huang, Xia

    2011-09-01

    A runoff retention and filtration pilot system was designed and the long-term purification effect of the runoff was monitored. Runoff pollution characters in 2 typical events and treatment effect of the pilot system were analyzed. The results showed that the runoff was severely polluted. Event mean concentrations (EMCs) of SS, COD, TN and TP in the runoff were 361, 135, 7.88 and 0.62 mg/L respectively. The runoff formed by long rain presented an obvious first flush effect. The first 25% flow contributed more than 50% of the total pollutants loading of SS, TP, DTP and PO4(3-). The pilot system could reduce 100% of the non-point source pollution if the volume of the runoff was less than the retention tank. Otherwise the overflow will be purification by the filtration pilot system and the removal rates of SS, COD, TN, TP, DTP and PO4(3-) reached 97.4% , 61.8%, 22.6%, 85.1%, 72.1%, and 85.2% respectively. The system was stable and the removal rate of SS, COD, TN, and TP were 98.6%, 65.4%, 55.1% and 92.6%. The whole system could effectively remove the non-point source pollution caused by runoff.

  16. Existing data sources for clinical epidemiology: the Danish Patient Compensation Association database

    Directory of Open Access Journals (Sweden)

    Tilma J

    2015-07-01

    Full Text Available Jens Tilma,1 Mette Nørgaard,1 Kim Lyngby Mikkelsen,2 Søren Paaske Johnsen1 1Department of Clinical Epidemiology, Aarhus University Hospital, Aarhus, 2Danish Patient Compensation Association, Copenhagen, Denmark Abstract: Any patient in the Danish health care system who experiences a treatment injury can make a compensation claim to the Danish Patient Compensation Association (DPCA free of charge. The aim of this paper is to describe the DPCA database as a source of data for epidemiological research. Data to DPCA are collected prospectively on all claims and include information on patient factors and health records, system factors, and administrative data. Approval of claims is based on injury due to the principle of treatment below experienced specialist standard or intolerable, unexpected extensiveness of injury. Average processing time of a compensation claim is 6–8 months. Data collection is nationwide and started in 1992. The patient's central registration system number, a unique personal identifier, allows for data linkage to other registries such as the Danish National Patient Registry. The DPCA data are accessible for research following data usage permission and make it possible to analyze all claims or specific subgroups to identify predictors, outcomes, etc. DPCA data have until now been used only in few studies but could be a useful data source in future studies of health care-related injuries. Keywords: public health care, treatment injuries, no-fault compensation, registries, research, Denmark

  17. Launching and controlling Gaussian beams from point sources via planar transformation media

    Science.gov (United States)

    Odabasi, Hayrettin; Sainath, Kamalesh; Teixeira, Fernando L.

    2018-02-01

    Based on operations prescribed under the paradigm of complex transformation optics (CTO) [F. Teixeira and W. Chew, J. Electromagn. Waves Appl. 13, 665 (1999), 10.1163/156939399X01104; F. L. Teixeira and W. C. Chew, Int. J. Numer. Model. 13, 441 (2000), 10.1002/1099-1204(200009/10)13:5%3C441::AID-JNM376%3E3.0.CO;2-J; H. Odabasi, F. L. Teixeira, and W. C. Chew, J. Opt. Soc. Am. B 28, 1317 (2011), 10.1364/JOSAB.28.001317; B.-I. Popa and S. A. Cummer, Phys. Rev. A 84, 063837 (2011), 10.1103/PhysRevA.84.063837], it was recently shown in [G. Castaldi, S. Savoia, V. Galdi, A. Alù, and N. Engheta, Phys. Rev. Lett. 110, 173901 (2013), 10.1103/PhysRevLett.110.173901] that a complex source point (CSP) can be mimicked by parity-time (PT ) transformation media. Such coordinate transformation has a mirror symmetry for the imaginary part, and results in a balanced loss/gain metamaterial slab. A CSP produces a Gaussian beam and, consequently, a point source placed at the center of such a metamaterial slab produces a Gaussian beam propagating away from the slab. Here, we extend the CTO analysis to nonsymmetric complex coordinate transformations as put forth in [S. Savoia, G. Castaldi, and V. Galdi, J. Opt. 18, 044027 (2016), 10.1088/2040-8978/18/4/044027] and verify that, by using simply a (homogeneous) doubly anisotropic gain-media metamaterial slab, one can still mimic a CSP and produce Gaussian beam. In addition, we show that a Gaussian-like beams can be produced by point sources placed outside the slab as well. By making use of the extra degrees of freedom (the real and imaginary parts of the coordinate transformation) provided by CTO, the near-zero requirement on the real part of the resulting constitutive parameters can be relaxed to facilitate potential realization of Gaussian-like beams. We illustrate how beam properties such as peak amplitude and waist location can be controlled by a proper choice of (complex-valued) CTO Jacobian elements. In particular, the beam waist

  18. Technical Note: A new global database of trace gases and aerosols from multiple sources of high vertical resolution measurements

    Directory of Open Access Journals (Sweden)

    G. E. Bodeker

    2008-09-01

    Full Text Available A new database of trace gases and aerosols with global coverage, derived from high vertical resolution profile measurements, has been assembled as a collection of binary data files; hereafter referred to as the "Binary DataBase of Profiles" (BDBP. Version 1.0 of the BDBP, described here, includes measurements from different satellite- (HALOE, POAM II and III, SAGE I and II and ground-based measurement systems (ozonesondes. In addition to the primary product of ozone, secondary measurements of other trace gases, aerosol extinction, and temperature are included. All data are subjected to very strict quality control and for every measurement a percentage error on the measurement is included. To facilitate analyses, each measurement is added to 3 different instances (3 different grids of the database where measurements are indexed by: (1 geographic latitude, longitude, altitude (in 1 km steps and time, (2 geographic latitude, longitude, pressure (at levels ~1 km apart and time, (3 equivalent latitude, potential temperature (8 levels from 300 K to 650 K and time.

    In contrast to existing zonal mean databases, by including a wider range of measurement sources (both satellite and ozonesondes, the BDBP is sufficiently dense to permit calculation of changes in ozone by latitude, longitude and altitude. In addition, by including other trace gases such as water vapour, this database can be used for comprehensive radiative transfer calculations. By providing the original measurements rather than derived monthly means, the BDBP is applicable to a wider range of applications than databases containing only monthly mean data. Monthly mean zonal mean ozone concentrations calculated from the BDBP are compared with the database of Randel and Wu, which has been used in many earlier analyses. As opposed to that database which is generated from regression model fits, the BDBP uses the original (quality controlled measurements with no smoothing applied in any

  19. Relative accuracy and availability of an Irish National Database of dispensed medication as a source of medication history information: observational study and retrospective record analysis.

    LENUS (Irish Health Repository)

    Grimes, T

    2013-01-27

    WHAT IS KNOWN AND OBJECTIVE: The medication reconciliation process begins by identifying which medicines a patient used before presentation to hospital. This is time-consuming, labour intensive and may involve interruption of clinicians. We sought to identify the availability and accuracy of data held in a national dispensing database, relative to other sources of medication history information. METHODS: For patients admitted to two acute hospitals in Ireland, a Gold Standard Pre-Admission Medication List (GSPAML) was identified and corroborated with the patient or carer. The GSPAML was compared for accuracy and availability to PAMLs from other sources, including the Health Service Executive Primary Care Reimbursement Scheme (HSE-PCRS) dispensing database. RESULTS: Some 1111 medication were assessed for 97 patients, who were median age 74 years (range 18-92 years), median four co-morbidities (range 1-9), used median 10 medications (range 3-25) and half (52%) were male. The HSE-PCRS PAML was the most accurate source compared to lists provided by the general practitioner, community pharmacist or cited in previous hospital documentation: the list agreed for 74% of the medications the patients actually used, representing complete agreement for all medications in 17% of patients. It was equally contemporaneous to other sources, but was less reliable for male than female patients, those using increasing numbers of medications and those using one or more item that was not reimbursable by the HSE. WHAT IS NEW AND CONCLUSION: The HSE-PCRS database is a relatively accurate, available and contemporaneous source of medication history information and could support acute hospital medication reconciliation.

  20. Impact of Point and Non-point Source Pollution on Coral Reef Ecosystems In Mamala Bay, Oahu, Hawaii based on Water Quality Measurements and Benthic Surveys in 1993-1994 (NODC Accession 0001172)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The effects of both point and non-point sources of pollution on coral reef ecosystems in Mamala Bay were studied at three levels of biological organization; the...

  1. BioWarehouse: a bioinformatics database warehouse toolkit.

    Science.gov (United States)

    Lee, Thomas J; Pouliot, Yannick; Wagner, Valerie; Gupta, Priyanka; Stringer-Calvert, David W J; Tenenbaum, Jessica D; Karp, Peter D

    2006-03-23

    This article addresses the problem of interoperation of heterogeneous bioinformatics databases. We introduce BioWarehouse, an open source toolkit for constructing bioinformatics database warehouses using the MySQL and Oracle relational database managers. BioWarehouse integrates its component databases into a common representational framework within a single database management system, thus enabling multi-database queries using the Structured Query Language (SQL) but also facilitating a variety of database integration tasks such as comparative analysis and data mining. BioWarehouse currently supports the integration of a pathway-centric set of databases including ENZYME, KEGG, and BioCyc, and in addition the UniProt, GenBank, NCBI Taxonomy, and CMR databases, and the Gene Ontology. Loader tools, written in the C and JAVA languages, parse and load these databases into a relational database schema. The loaders also apply a degree of semantic normalization to their respective source data, decreasing semantic heterogeneity. The schema supports the following bioinformatics datatypes: chemical compounds, biochemical reactions, metabolic pathways, proteins, genes, nucleic acid sequences, features on protein and nucleic-acid sequences, organisms, organism taxonomies, and controlled vocabularies. As an application example, we applied BioWarehouse to determine the fraction of biochemically characterized enzyme activities for which no sequences exist in the public sequence databases. The answer is that no sequence exists for 36% of enzyme activities for which EC numbers have been assigned. These gaps in sequence data significantly limit the accuracy of genome annotation and metabolic pathway prediction, and are a barrier for metabolic engineering. Complex queries of this type provide examples of the value of the data warehousing approach to bioinformatics research. BioWarehouse embodies significant progress on the database integration problem for bioinformatics.

  2. History and Point in Time in Enterprise Applications

    Directory of Open Access Journals (Sweden)

    Constantin Gelu APOSTOL

    2006-01-01

    Full Text Available First part points out the main differences between temporal and non-temporal databases. In the second part, based on identification of the three main categories of time involved in database applications: user-defined time, valid time and transaction time, some relevant solutions for their implementation are discussed, mainly from the point of view of database organization and data access level of enterprise applications. The final part is dedicated to the influences of historical data in the business logic and presentation levels of enterprise applications and in application services, as security, workflow, reporting.

  3. Food Habits Database (FHDBS)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The NEFSC Food Habits Database has two major sources of data. The first, and most extensive, is the standard NEFSC Bottom Trawl Surveys Program. During these...

  4. Extraction of Point Source Gamma Signals from Aerial Survey Data Taken over a Las Vegas Nevada, Residential Area

    International Nuclear Information System (INIS)

    Thane J. Hendricks

    2007-01-01

    Detection of point-source gamma signals from aerial measurements is complicated by widely varying terrestrial gamma backgrounds, since these variations frequently resemble signals from point-sources. Spectral stripping techniques have been very useful in separating man-made and natural radiation contributions which exist on Energy Research and Development Administration (ERDA) plant sites and other like facilities. However, these facilities are generally situated in desert areas or otherwise flat terrain with few man-made structures to disturb the natural background. It is of great interest to determine if the stripping technique can be successfully applied in populated areas where numerous man-made disturbances (houses, streets, yards, vehicles, etc.) exist

  5. Study on the quantitative relationship between Agricultural water and fertilization process and non-point source pollution based on field experiments

    Science.gov (United States)

    Wang, H.; Chen, K.; Wu, Z.; Guan, X.

    2017-12-01

    In recent years, with the prominent of water environment problem and the relative increase of point source pollution governance, especially the agricultural non-point source pollution problem caused by the extensive use of fertilizers and pesticides has become increasingly aroused people's concern and attention. In order to reveal the quantitative relationship between agriculture water and fertilizer and non-point source pollution, on the basis of elm field experiment and combined with agricultural drainage irrigation model, the agricultural irrigation water and the relationship between fertilizer and fertilization scheme and non-point source pollution were analyzed and calculated by field emission intensity index. The results show that the variation of displacement varies greatly under different irrigation conditions. When the irrigation water increased from 22cm to 42cm, the irrigation water increased by 20 cm while the field displacement increased by 11.92 cm, about 66.22% of the added value of irrigation water. Then the irrigation water increased from 42 to 68, irrigation water increased 26 cm, and the field displacement increased by 22.48 cm, accounting for 86.46% of irrigation water. So there is an "inflection point" between the irrigation water amount and field displacement amount. The load intensity increases with the increase of irrigation water and shows a significant power correlation. Under the different irrigation condition, the increase amplitude of load intensity with the increase of irrigation water is different. When the irrigation water is smaller, the load intensity increase relatively less, and when the irrigation water increased to about 42 cm, the load intensity will increase considerably. In addition, there was a positive correlation between the fertilization and load intensity. The load intensity had obvious difference in different fertilization modes even with same fertilization level, in which the fertilizer field unit load intensity

  6. The rise in the positron fraction. Distance limits on positron point sources from cosmic ray arrival directions and diffuse gamma-rays

    Energy Technology Data Exchange (ETDEWEB)

    Gebauer, Iris; Bentele, Rosemarie [Karlsruhe Institute of Technology, Karlsruhe (Germany)

    2016-07-01

    The rise in the positron fraction as observed by AMS and previously by PAMELA, cannot be explained by the standard paradigm of cosmic ray transport in which positrons are produced by cosmic-ray-gas interactions in the interstellar medium. Possible explanations are pulsars, which produce energetic electron-positron pairs in their rotating magnetic fields, or the annihilation of dark matter. Here we assume that these positrons originate from a single close-by point source, producing equal amounts of electrons and positrons. The propagation and energy losses of these electrons and positrons are calculated numerically using the DRAGON code, the source properties are optimized to best describe the AMS data. Using the FERMI-LAT limits on a possible dipole anisotropy in electron and positron arrival directions, we put a limit on the minimum distance of such a point source. The energy losses that these energetic electrons and positrons suffer on their way through the galaxy create gamma ray photons through bremsstrahlung and Inverse Compton scattering. Using the measurement of diffuse gamma rays from Fermi-LAT we put a limit on the maximum distance of such a point source. We find that a single electron positron point source powerful enough to explain the locally observed positron fraction must reside between 225 pc and 3.7 kpc distance from the sun and compare to known pulsars.

  7. Directory of IAEA databases. 4. ed.

    International Nuclear Information System (INIS)

    1997-06-01

    This fourth edition of the Directory of IAEA Databases has been prepared within the Division of NESI. ITs main objective is to describe the computerized information sources available to the public. This directory contains all publicly available databases which are produced at the IAEA. This includes databases stored on the mainframe, LAN servers and user PCs. All IAEA Division Directors have been requested to register the existence of their databases with NESI. At the data of printing, some of the information in the directory will be already obsolete. For the most up-to-date information please see the IAEA's World Wide Web site at URL: http:/www.iaea.or.at/databases/dbdir/. Refs, figs, tabs

  8. Palaeo sea-level and ice-sheet databases: problems, strategies and perspectives

    Science.gov (United States)

    Rovere, Alessio; Düsterhus, André; Carlson, Anders; Barlow, Natasha; Bradwell, Tom; Dutton, Andrea; Gehrels, Roland; Hibbert, Fiona; Hijma, Marc; Horton, Benjamin; Klemann, Volker; Kopp, Robert; Sivan, Dorit; Tarasov, Lev; Törnqvist, Torbjorn

    2016-04-01

    Databases of palaeoclimate data have driven many major developments in understanding the Earth system. The measurement and interpretation of palaeo sea-level and ice-sheet data that form such databases pose considerable challenges to the scientific communities that use them for further analyses. In this paper, we build on the experience of the PALSEA (PALeo constraints on SEA level rise) community, which is a working group inside the PAGES (Past Global Changes) project, to describe the challenges and best strategies that can be adopted to build a self-consistent and standardised database of geological and geochemical data related to palaeo sea levels and ice sheets. Our aim in this paper is to identify key points that need attention and subsequent funding when undertaking the task of database creation. We conclude that any sea-level or ice-sheet database must be divided into three instances: i) measurement; ii) interpretation; iii) database creation. Measurement should include postion, age, description of geological features, and quantification of uncertainties. All must be described as objectively as possible. Interpretation can be subjective, but it should always include uncertainties and include all the possible interpretations, without unjustified a priori exclusions. We propose that, in the creation of a database, an approach based on Accessibility, Transparency, Trust, Availability, Continued updating, Completeness and Communication of content (ATTAC3) must be adopted. Also, it is essential to consider the community structure that creates and benefits of a database. We conclude that funding sources should consider to address not only the creation of original data in specific research-question oriented projects, but also include the possibility to use part of the funding for IT-related and database creation tasks, which are essential to guarantee accessibility and maintenance of the collected data.

  9. BioWarehouse: a bioinformatics database warehouse toolkit

    Directory of Open Access Journals (Sweden)

    Stringer-Calvert David WJ

    2006-03-01

    Full Text Available Abstract Background This article addresses the problem of interoperation of heterogeneous bioinformatics databases. Results We introduce BioWarehouse, an open source toolkit for constructing bioinformatics database warehouses using the MySQL and Oracle relational database managers. BioWarehouse integrates its component databases into a common representational framework within a single database management system, thus enabling multi-database queries using the Structured Query Language (SQL but also facilitating a variety of database integration tasks such as comparative analysis and data mining. BioWarehouse currently supports the integration of a pathway-centric set of databases including ENZYME, KEGG, and BioCyc, and in addition the UniProt, GenBank, NCBI Taxonomy, and CMR databases, and the Gene Ontology. Loader tools, written in the C and JAVA languages, parse and load these databases into a relational database schema. The loaders also apply a degree of semantic normalization to their respective source data, decreasing semantic heterogeneity. The schema supports the following bioinformatics datatypes: chemical compounds, biochemical reactions, metabolic pathways, proteins, genes, nucleic acid sequences, features on protein and nucleic-acid sequences, organisms, organism taxonomies, and controlled vocabularies. As an application example, we applied BioWarehouse to determine the fraction of biochemically characterized enzyme activities for which no sequences exist in the public sequence databases. The answer is that no sequence exists for 36% of enzyme activities for which EC numbers have been assigned. These gaps in sequence data significantly limit the accuracy of genome annotation and metabolic pathway prediction, and are a barrier for metabolic engineering. Complex queries of this type provide examples of the value of the data warehousing approach to bioinformatics research. Conclusion BioWarehouse embodies significant progress on the

  10. CHANDRA ACIS SURVEY OF X-RAY POINT SOURCES IN NEARBY GALAXIES. II. X-RAY LUMINOSITY FUNCTIONS AND ULTRALUMINOUS X-RAY SOURCES

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Song; Qiu, Yanli; Liu, Jifeng [Key Laboratory of Optical Astronomy, National Astronomical Observatories, Chinese Academy of Sciences, Beijing 100012 (China); Bregman, Joel N., E-mail: songw@bao.ac.cn, E-mail: jfliu@bao.ac.cn [University of Michigan, Ann Arbor, MI 48109 (United States)

    2016-09-20

    Based on the recently completed Chandra /ACIS survey of X-ray point sources in nearby galaxies, we study the X-ray luminosity functions (XLFs) for X-ray point sources in different types of galaxies and the statistical properties of ultraluminous X-ray sources (ULXs). Uniform procedures are developed to compute the detection threshold, to estimate the foreground/background contamination, and to calculate the XLFs for individual galaxies and groups of galaxies, resulting in an XLF library of 343 galaxies of different types. With the large number of surveyed galaxies, we have studied the XLFs and ULX properties across different host galaxy types, and confirm with good statistics that the XLF slope flattens from lenticular ( α ∼ 1.50 ± 0.07) to elliptical (∼1.21 ± 0.02), to spirals (∼0.80 ± 0.02), to peculiars (∼0.55 ± 0.30), and to irregulars (∼0.26 ± 0.10). The XLF break dividing the neutron star and black hole binaries is also confirmed, albeit at quite different break luminosities for different types of galaxies. A radial dependency is found for ellipticals, with a flatter XLF slope for sources located between D {sub 25} and 2 D {sub 25}, suggesting the XLF slopes in the outer region of early-type galaxies are dominated by low-mass X-ray binaries in globular clusters. This study shows that the ULX rate in early-type galaxies is 0.24 ± 0.05 ULXs per surveyed galaxy, on a 5 σ confidence level. The XLF for ULXs in late-type galaxies extends smoothly until it drops abruptly around 4 × 10{sup 40} erg s{sup −1}, and this break may suggest a mild boundary between the stellar black hole population possibly including 30 M {sub ⊙} black holes with super-Eddington radiation and intermediate mass black holes.

  11. Reliability databases: State-of-the-art and perspectives

    DEFF Research Database (Denmark)

    Akhmedjanov, Farit

    2001-01-01

    The report gives a history of development and an overview of the existing reliability databases. This overview also describes some other (than computer databases) sources of reliability and failures information, e.g. reliability handbooks, but the mainattention is paid to standard models...... and software packages containing the data mentioned. The standards corresponding to collection and exchange of reliability data are observed too. Finally, perspective directions in such data sources development areshown....

  12. The plant phenological online database (PPODB): an online database for long-term phenological data

    Science.gov (United States)

    Dierenbach, Jonas; Badeck, Franz-W.; Schaber, Jörg

    2013-09-01

    We present an online database that provides unrestricted and free access to over 16 million plant phenological observations from over 8,000 stations in Central Europe between the years 1880 and 2009. Unique features are (1) a flexible and unrestricted access to a full-fledged database, allowing for a wide range of individual queries and data retrieval, (2) historical data for Germany before 1951 ranging back to 1880, and (3) more than 480 curated long-term time series covering more than 100 years for individual phenological phases and plants combined over Natural Regions in Germany. Time series for single stations or Natural Regions can be accessed through a user-friendly graphical geo-referenced interface. The joint databases made available with the plant phenological database PPODB render accessible an important data source for further analyses of long-term changes in phenology. The database can be accessed via www.ppodb.de .

  13. Database theory and SQL practice using Access

    International Nuclear Information System (INIS)

    Kim, Gyeong Min; Lee, Myeong Jin

    2001-01-01

    This book introduces database theory and SQL practice using Access. It is comprised of seven chapters, which give description of understanding database with basic conception and DMBS, understanding relational database with examples of it, building database table and inputting data using access 2000, structured Query Language with introduction, management and making complex query using SQL, command for advanced SQL with understanding conception of join and virtual table, design on database for online bookstore with six steps and building of application with function, structure, component, understanding of the principle, operation and checking programming source for application menu.

  14. Morphology, chemistry and distribution of neoformed spherulites in agricultural land affected by metallurgical point-source pollution

    NARCIS (Netherlands)

    Leguedois, S.; Oort, van F.; Jongmans, A.G.; Chevalier, P.

    2004-01-01

    Metal distribution patterns in superficial soil horizons of agricultural land affected by metallurgical point-source pollution were studied using optical and electron microscopy, synchrotron radiation and spectroscopy analyses. The site is located in northern France, at the center of a former entry

  15. Identification and quantification of point sources of surface water contamination in fruit culture in the Netherlands

    NARCIS (Netherlands)

    Wenneker, M.; Beltman, W.H.J.; Werd, de H.A.E.; Zande, van de J.C.

    2008-01-01

    Measurements of pesticide concentrations in surface water by the water boards show that they have decreased less than was expected from model calculations. Possibly, the implementation of spray drift reducing techniques is overestimated in the model calculation. The impact of point sources is

  16. Estimation of Methane Emissions from Municipal Solid Waste Landfills in China Based on Point Emission Sources

    Directory of Open Access Journals (Sweden)

    Cai Bo-Feng

    2014-01-01

    Citation: Cai, B.-F., Liu, J.-G., Gao, Q.-X., et al., 2014. Estimation of methane emissions from municipal solid waste landfills in China based on point emission sources. Adv. Clim. Change Res. 5(2, doi: 10.3724/SP.J.1248.2014.081.

  17. VizieR Online Data Catalog: ChaMP X-ray point source catalog (Kim+, 2007)

    Science.gov (United States)

    Kim, M.; Kim, D.-W.; Wilkes, B. J.; Green, P. J.; Kim, E.; Anderson, C. S.; Barkhouse, W. A.; Evans, N. R.; Ivezic, Z.; Karovska, M.; Kashyap, V. L.; Lee, M. G.; Maksym, P.; Mossman, A. E.; Silverman, J. D.; Tananbaum, H. D.

    2009-01-01

    We present the Chandra Multiwavelength Project (ChaMP) X-ray point source catalog with ~6800 X-ray sources detected in 149 Chandra observations covering ~10deg2. The full ChaMP catalog sample is 7 times larger than the initial published ChaMP catalog. The exposure time of the fields in our sample ranges from 0.9 to 124ks, corresponding to a deepest X-ray flux limit of f0.5-8.0=9x10-16ergs/cm2/s. The ChaMP X-ray data have been uniformly reduced and analyzed with ChaMP-specific pipelines and then carefully validated by visual inspection. The ChaMP catalog includes X-ray photometric data in eight different energy bands as well as X-ray spectral hardness ratios and colors. To best utilize the ChaMP catalog, we also present the source reliability, detection probability, and positional uncertainty. (10 data files).

  18. Multi-angle Indicators System of Non-point Pollution Source Assessment in Rural Areas: A Case Study Near Taihu Lake

    Science.gov (United States)

    Huang, Lei; Ban, Jie; Han, Yu Ting; Yang, Jie; Bi, Jun

    2013-04-01

    This study aims to identify key environmental risk sources contributing to water eutrophication and to suggest certain risk management strategies for rural areas. The multi-angle indicators included in the risk source assessment system were non-point source pollution, deficient waste treatment, and public awareness of environmental risk, which combined psychometric paradigm methods, the contingent valuation method, and personal interviews to describe the environmental sensitivity of local residents. Total risk values of different villages near Taihu Lake were calculated in the case study, which resulted in a geographic risk map showing which village was the critical risk source of Taihu eutrophication. The increased application of phosphorus (P) and nitrogen (N), loss vulnerability of pollutant, and a lack of environmental risk awareness led to more serious non-point pollution, especially in rural China. Interesting results revealed by the quotient between the scores of objective risk sources and subjective risk sources showed what should be improved for each study village. More environmental investments, control of agricultural activities, and promotion of environmental education are critical considerations for rural environmental management. These findings are helpful for developing targeted and effective risk management strategies in rural areas.

  19. Carbon dioxide capture and separation techniques for advanced power generation point sources

    Energy Technology Data Exchange (ETDEWEB)

    Pennline, H.W.; Luebke, D.R.; Morsi, B.I.; Heintz, Y.J.; Jones, K.L.; Ilconich, J.B.

    2006-09-01

    The capture/separation step for carbon dioxide (CO2) from large-point sources is a critical one with respect to the technical feasibility and cost of the overall carbon sequestration scenario. For large-point sources, such as those found in power generation, the carbon dioxide capture techniques being investigated by the in-house research area of the National Energy Technology Laboratory possess the potential for improved efficiency and costs as compared to more conventional technologies. The investigated techniques can have wide applications, but the research has focused on capture/separation of carbon dioxide from flue gas (postcombustion from fossil fuel-fired combustors) and from fuel gas (precombustion, such as integrated gasification combined cycle – IGCC). With respect to fuel gas applications, novel concepts are being developed in wet scrubbing with physical absorption; chemical absorption with solid sorbents; and separation by membranes. In one concept, a wet scrubbing technique is being investigated that uses a physical solvent process to remove CO2 from fuel gas of an IGCC system at elevated temperature and pressure. The need to define an ideal solvent has led to the study of the solubility and mass transfer properties of various solvents. Fabrication techniques and mechanistic studies for hybrid membranes separating CO2 from the fuel gas produced by coal gasification are also being performed. Membranes that consist of CO2-philic silanes incorporated into an alumina support or ionic liquids encapsulated into a polymeric substrate have been investigated for permeability and selectivity. An overview of two novel techniques is presented along with a research progress status of each technology.

  20. SIMS: addressing the problem of heterogeneity in databases

    Science.gov (United States)

    Arens, Yigal

    1997-02-01

    The heterogeneity of remotely accessible databases -- with respect to contents, query language, semantics, organization, etc. -- presents serious obstacles to convenient querying. The SIMS (single interface to multiple sources) system addresses this global integration problem. It does so by defining a single language for describing the domain about which information is stored in the databases and using this language as the query language. Each database to which SIMS is to provide access is modeled using this language. The model describes a database's contents, organization, and other relevant features. SIMS uses these models, together with a planning system drawing on techniques from artificial intelligence, to decompose a given user's high-level query into a series of queries against the databases and other data manipulation steps. The retrieval plan is constructed so as to minimize data movement over the network and maximize parallelism to increase execution speed. SIMS can recover from network failures during plan execution by obtaining data from alternate sources, when possible. SIMS has been demonstrated in the domains of medical informatics and logistics, using real databases.

  1. Domain Regeneration for Cross-Database Micro-Expression Recognition

    Science.gov (United States)

    Zong, Yuan; Zheng, Wenming; Huang, Xiaohua; Shi, Jingang; Cui, Zhen; Zhao, Guoying

    2018-05-01

    In this paper, we investigate the cross-database micro-expression recognition problem, where the training and testing samples are from two different micro-expression databases. Under this setting, the training and testing samples would have different feature distributions and hence the performance of most existing micro-expression recognition methods may decrease greatly. To solve this problem, we propose a simple yet effective method called Target Sample Re-Generator (TSRG) in this paper. By using TSRG, we are able to re-generate the samples from target micro-expression database and the re-generated target samples would share same or similar feature distributions with the original source samples. For this reason, we can then use the classifier learned based on the labeled source samples to accurately predict the micro-expression categories of the unlabeled target samples. To evaluate the performance of the proposed TSRG method, extensive cross-database micro-expression recognition experiments designed based on SMIC and CASME II databases are conducted. Compared with recent state-of-the-art cross-database emotion recognition methods, the proposed TSRG achieves more promising results.

  2. Database Search Engines: Paradigms, Challenges and Solutions.

    Science.gov (United States)

    Verheggen, Kenneth; Martens, Lennart; Berven, Frode S; Barsnes, Harald; Vaudel, Marc

    2016-01-01

    The first step in identifying proteins from mass spectrometry based shotgun proteomics data is to infer peptides from tandem mass spectra, a task generally achieved using database search engines. In this chapter, the basic principles of database search engines are introduced with a focus on open source software, and the use of database search engines is demonstrated using the freely available SearchGUI interface. This chapter also discusses how to tackle general issues related to sequence database searching and shows how to minimize their impact.

  3. Directory of IAEA databases. 3. ed.

    International Nuclear Information System (INIS)

    1993-12-01

    This second edition of the Directory of IAEA Databases has been prepared within the Division of Scientific and Technical Information. Its main objective is to describe the computerized information sources available to staff members. This directory contains all databases produced at the IAEA, including databases stored on the mainframe, LAN's and PC's. All IAEA Division Directors have been requested to register the existence of their databases with NESI. For the second edition database owners were requested to review the existing entries for their databases and answer four additional questions. The four additional questions concerned the type of database (e.g. Bibliographic, Text, Statistical etc.), the category of database (e.g. Administrative, Nuclear Data etc.), the available documentation and the type of media used for distribution. In the individual entries on the following pages the answers to the first two questions (type and category) is always listed, but the answer to the second two questions (documentation and media) is only listed when information has been made available

  4. Move Over, Word Processors--Here Come the Databases.

    Science.gov (United States)

    Olds, Henry F., Jr.; Dickenson, Anne

    1985-01-01

    Discusses the use of beginning, intermediate, and advanced databases for instructional purposes. A table listing seven databases with information on ease of use, smoothness of operation, data capacity, speed, source, and program features is included. (JN)

  5. Emissions of perfluorinated alkylated substances (PFAS) from point sources--identification of relevant branches.

    Science.gov (United States)

    Clara, M; Scheffknecht, C; Scharf, S; Weiss, S; Gans, O

    2008-01-01

    Effluents of wastewater treatment plants are relevant point sources for the emission of hazardous xenobiotic substances to the aquatic environment. One group of substances, which recently entered scientific and political discussions, is the group of the perfluorinated alkylated substances (PFAS). The most studied compounds from this group are perfluorooctanoic acid (PFOA) and perfluorooctane sulphonate (PFOS), which are the most important degradation products of PFAS. These two substances are known to be persistent, bioaccumulative and toxic (PBT). In the present study, eleven PFAS were investigated in effluents of municipal wastewater treatment plants (WWTP) and in industrial wastewaters. PFOS and PFOA proved to be the dominant compounds in all sampled wastewaters. Concentrations of up to 340 ng/L of PFOS and up to 220 ng/L of PFOA were observed. Besides these two compounds, perfluorohexanoic acid (PFHxA) was also present in nearly all effluents and maximum concentrations of up to 280 ng/L were measured. Only N-ethylperfluorooctane sulphonamide (N-EtPFOSA) and its degradation/metabolisation product perfluorooctane sulphonamide (PFOSA) were either detected below the limit of quantification or were not even detected at all. Beside the effluents of the municipal WWTPs, nine industrial wastewaters from six different industrial branches were also investigated. Significantly, the highest emissions or PFOS were observed from metal industry whereas paper industry showed the highest PFOA emission. Several PFAS, especially perfluorononanoic acid (PFNA), perfluorodecanoic acid (PFDA), perfluorododecanoic acid (PFDoA) and PFOS are predominantly emitted from industrial sources, with concentrations being a factor of 10 higher than those observed in the municipal WWTP effluents. Perfluorodecane sulphonate (PFDS), N-Et-PFOSA and PFOSA were not detected in any of the sampled industrial point sources. (c) IWA Publishing 2008.

  6. Column Number Density Expressions Through M = 0 and M = 1 Point Source Plumes Along Any Straight Path

    Science.gov (United States)

    Woronowicz, Michael

    2016-01-01

    Analytical expressions for column number density (CND) are developed for optical line of sight paths through a variety of steady free molecule point source models including directionally-constrained effusion (Mach number M = 0) and flow from a sonic orifice (M = 1). Sonic orifice solutions are approximate, developed using a fair simulacrum fitted to the free molecule solution. Expressions are also developed for a spherically-symmetric thermal expansion (M = 0). CND solutions are found for the most general paths relative to these sources and briefly explored. It is determined that the maximum CND from a distant location through directed effusion and sonic orifice cases occurs along the path parallel to the source plane that intersects the plume axis. For the effusive case this value is exactly twice the CND found along the ray originating from that point of intersection and extending to infinity along the plume's axis. For sonic plumes this ratio is reduced to about 4/3. For high Mach number cases the maximum CND will be found along the axial centerline path. Keywords: column number density, plume flows, outgassing, free molecule flow.

  7. Low energy electron point source microscopy: beyond imaging

    Energy Technology Data Exchange (ETDEWEB)

    Beyer, Andre; Goelzhaeuser, Armin [Physics of Supramolecular Systems and Surfaces, University of Bielefeld, Postfach 100131, 33501 Bielefeld (Germany)

    2010-09-01

    Low energy electron point source (LEEPS) microscopy has the capability to record in-line holograms at very high magnifications with a fairly simple set-up. After the holograms are numerically reconstructed, structural features with the size of about 2 nm can be resolved. The achievement of an even higher resolution has been predicted. However, a number of obstacles are known to impede the realization of this goal, for example the presence of electric fields around the imaged object, electrostatic charging or radiation induced processes. This topical review gives an overview of the achievements as well as the difficulties in the efforts to shift the resolution limit of LEEPS microscopy towards the atomic level. A special emphasis is laid on the high sensitivity of low energy electrons to electrical fields, which limits the structural determination of the imaged objects. On the other hand, the investigation of the electrical field around objects of known structure is very useful for other tasks and LEEPS microscopy can be extended beyond the task of imaging. The determination of the electrical resistance of individual nanowires can be achieved by a proper analysis of the corresponding LEEPS micrographs. This conductivity imaging may be a very useful application for LEEPS microscopes. (topical review)

  8. The RHIC transfer line cable database

    International Nuclear Information System (INIS)

    Scholl, E.H.; Satogata, T.

    1995-01-01

    A cable database was created to facilitate and document installation of cables and wiring in the RHIC project, as well as to provide a data source to track possible wiring and signal problems. The eight tables of this relational database, currently implemented in Sybase, contain information ranging from cable routing to attenuation of individual wires. This database was created in a hierarchical scheme under the assumption that cables contain wires -- each instance of a cable has one to many wires associated with it. This scheme allows entry of information pertinent to individual wires while only requiring single entries for each cable. Relationships to other RHIC databases are also discussed

  9. Presentation layer finding database of cyanobacteria and algae

    OpenAIRE

    SEMECKÝ, Jiří

    2012-01-01

    Phycological Laboratory University of South Bohemia in Czech Budejovice uses occurrence database samples. This work deals with the analysis and optimization of the existing database, designing and programming extension that allows processing points based on GPS coordinates and display them in on-line maps and georeferenced image.

  10. Database of Interacting Proteins (DIP)

    Data.gov (United States)

    U.S. Department of Health & Human Services — The DIP database catalogs experimentally determined interactions between proteins. It combines information from a variety of sources to create a single, consistent...

  11. Normalized Point Source Sensitivity for Off-Axis Optical Performance Evaluation of the Thirty Meter Telescope

    Science.gov (United States)

    Seo, Byoung-Joon; Nissly, Carl; Troy, Mitchell; Angeli, George

    2010-01-01

    The Normalized Point Source Sensitivity (PSSN) has previously been defined and analyzed as an On-Axis seeing-limited telescope performance metric. In this paper, we expand the scope of the PSSN definition to include Off-Axis field of view (FoV) points and apply this generalized metric for performance evaluation of the Thirty Meter Telescope (TMT). We first propose various possible choices for the PSSN definition and select one as our baseline. We show that our baseline metric has useful properties including the multiplicative feature even when considering Off-Axis FoV points, which has proven to be useful for optimizing the telescope error budget. Various TMT optical errors are considered for the performance evaluation including segment alignment and phasing, segment surface figures, temperature, and gravity, whose On-Axis PSSN values have previously been published by our group.

  12. ALFRED: An Allele Frequency Database for Microevolutionary Studies

    Directory of Open Access Journals (Sweden)

    Kenneth K Kidd

    2005-01-01

    Full Text Available Many kinds of microevolutionary studies require data on multiple polymorphisms in multiple populations. Increasingly, and especially for human populations, multiple research groups collect relevant data and those data are dispersed widely in the literature. ALFRED has been designed to hold data from many sources and make them available over the web. Data are assembled from multiple sources, curated, and entered into the database. Multiple links to other resources are also established by the curators. A variety of search options are available and additional geographic based interfaces are being developed. The database can serve the human anthropologic genetic community by identifying what loci are already typed on many populations thereby helping to focus efforts on a common set of markers. The database can also serve as a model for databases handling similar DNA polymorphism data for other species.

  13. Comment on the internal consistency of thermodynamic databases supporting repository safety assessments

    International Nuclear Information System (INIS)

    Arthur, R.C.

    2001-11-01

    This report addresses the concept of internal consistency and its relevance to the reliability of thermodynamic databases used in repository safety assessments. In addition to being internally consistent, a reliable database should be accurate over a range of relevant temperatures and pressures, complete in the sense that all important aqueous species, gases and solid phases are represented, and traceable to original experimental results. No single definition of internal consistency need to be universally accepted as the most appropriate under all conditions, however. As a result, two databases that are each internally consistent may be inconsistent with respect to each other, and a database derived from two or more such databases must itself be internally inconsistent. The consequences of alternative definitions that are reasonably attributable to the concept of internal consistency can be illustrated with reference to the thermodynamic database supporting SKB's recent SR 97 safety assessment. This database is internally inconsistent because it includes equilibrium constants calculated over a range of temperatures: using conflicting reference values for some solids, gases and aqueous species that are common to two internally consistent databases (the OECD/NEA database for radioelements and SUPCRT databases for non-radioactive elements) that serve as source databases for the SR 97 TDB, using different definitions in these source databases of standard states for condensed phases and aqueous species, based on different mathematical expressions used in these source databases representing the temperature dependence of the heat capacity, and based on different chemical models adopted in these source databases for the aqueous phase. The importance of such inconsistencies must be considered in relation to the other database reliability criteria noted above, however. Thus, accepting a certain level of internal inconsistency in a database it is probably preferable to use a

  14. Comment on the internal consistency of thermodynamic databases supporting repository safety assessments

    Energy Technology Data Exchange (ETDEWEB)

    Arthur, R.C. [Monitor Scientific, LLC, Denver, CO (United States)

    2001-11-01

    This report addresses the concept of internal consistency and its relevance to the reliability of thermodynamic databases used in repository safety assessments. In addition to being internally consistent, a reliable database should be accurate over a range of relevant temperatures and pressures, complete in the sense that all important aqueous species, gases and solid phases are represented, and traceable to original experimental results. No single definition of internal consistency need to be universally accepted as the most appropriate under all conditions, however. As a result, two databases that are each internally consistent may be inconsistent with respect to each other, and a database derived from two or more such databases must itself be internally inconsistent. The consequences of alternative definitions that are reasonably attributable to the concept of internal consistency can be illustrated with reference to the thermodynamic database supporting SKB's recent SR 97 safety assessment. This database is internally inconsistent because it includes equilibrium constants calculated over a range of temperatures: using conflicting reference values for some solids, gases and aqueous species that are common to two internally consistent databases (the OECD/NEA database for radioelements and SUPCRT databases for non-radioactive elements) that serve as source databases for the SR 97 TDB, using different definitions in these source databases of standard states for condensed phases and aqueous species, based on different mathematical expressions used in these source databases representing the temperature dependence of the heat capacity, and based on different chemical models adopted in these source databases for the aqueous phase. The importance of such inconsistencies must be considered in relation to the other database reliability criteria noted above, however. Thus, accepting a certain level of internal inconsistency in a database it is probably preferable to

  15. Point-source reconstruction with a sparse light-sensor array for optical TPC readout

    International Nuclear Information System (INIS)

    Rutter, G; Richards, M; Bennieston, A J; Ramachers, Y A

    2011-01-01

    A reconstruction technique for sparse array optical signal readout is introduced and applied to the generic challenge of large-area readout of a large number of point light sources. This challenge finds a prominent example in future, large volume neutrino detector studies based on liquid argon. It is concluded that the sparse array option may be ruled out for reasons of required number of channels when compared to a benchmark derived from charge readout on wire-planes. Smaller-scale detectors, however, could benefit from this technology.

  16. Developing of database on nuclear power engineering and purchase of ORACLE system

    International Nuclear Information System (INIS)

    Liu Renkang

    1996-01-01

    This paper presents a point of view according development of database on the nuclear power engineering and performance of ORACLE database manager system. ORACLE system is a practical database system for purchasing

  17. Using E-PRTR data on point source emissions to air and water—First steps towards a national chemical footprint

    International Nuclear Information System (INIS)

    Sörme, L.; Palm, V.; Finnveden, G.

    2016-01-01

    There is a great need for indicators to monitor the use and potential impacts of hazardous chemicals. Today there is a huge lack of data, methods and results and method development and studies should be given urgent priority. The aim of this paper was to develop and test an approach to calculate the potential environmental impacts of chemicals for a whole country using the E-PRTR (European Pollutant Release and Transfer Register) as a database and Sweden as an example. Swedish data from 2008 on emissions to air and water for 54 substances from point sources were retrieved from an open database. The data were transformed and aggregated using USEtox, a life-cycle impact assessment (LCIA) method for calculating potential human toxicity and ecotoxicity, both from industrial emissions directly and after input–output analysis (IO analysis) to reallocate emissions to product categories. Zinc to air and water contributed most to human toxicity followed by mercury to air. The largest contribution by industry to potential human toxicity came from the metal industry, followed by the paper and paper product industry. For potential ecotoxicity, zinc, fluoranthene and copper contributed the most. The largest contributions by industry came from the paper and paper products manufacturing sector, followed by the basic metals manufacturing sector. The approach used here can be seen as the first step towards a chemical footprint for nations. By adding data from other countries and other sources, a more complete picture can be gained in line with other footprint calculations. Furthermore, diffuse emissions from, for example, transport or emissions of pesticides could also be added for a more holistic assessment. Since the area of chemicals is complicated, it is probably necessary to develop and use several indicators that complement each other. It is suggested that the approach outlined here could be useful in developing a method for establishing a national chemical footprint

  18. Using E-PRTR data on point source emissions to air and water—First steps towards a national chemical footprint

    Energy Technology Data Exchange (ETDEWEB)

    Sörme, L., E-mail: louise.sorme@scb.se [Statistics Sweden, Box 24300, SE-104 51 Sweden (Sweden); Palm, V. [Statistics Sweden, Box 24300, SE-104 51 Sweden (Sweden); KTH Royal Institute of Technology, Division of Environmental Strategies Research, SE-100 44 Stockholm (Sweden); Finnveden, G. [KTH Royal Institute of Technology, Division of Environmental Strategies Research, SE-100 44 Stockholm (Sweden)

    2016-01-15

    There is a great need for indicators to monitor the use and potential impacts of hazardous chemicals. Today there is a huge lack of data, methods and results and method development and studies should be given urgent priority. The aim of this paper was to develop and test an approach to calculate the potential environmental impacts of chemicals for a whole country using the E-PRTR (European Pollutant Release and Transfer Register) as a database and Sweden as an example. Swedish data from 2008 on emissions to air and water for 54 substances from point sources were retrieved from an open database. The data were transformed and aggregated using USEtox, a life-cycle impact assessment (LCIA) method for calculating potential human toxicity and ecotoxicity, both from industrial emissions directly and after input–output analysis (IO analysis) to reallocate emissions to product categories. Zinc to air and water contributed most to human toxicity followed by mercury to air. The largest contribution by industry to potential human toxicity came from the metal industry, followed by the paper and paper product industry. For potential ecotoxicity, zinc, fluoranthene and copper contributed the most. The largest contributions by industry came from the paper and paper products manufacturing sector, followed by the basic metals manufacturing sector. The approach used here can be seen as the first step towards a chemical footprint for nations. By adding data from other countries and other sources, a more complete picture can be gained in line with other footprint calculations. Furthermore, diffuse emissions from, for example, transport or emissions of pesticides could also be added for a more holistic assessment. Since the area of chemicals is complicated, it is probably necessary to develop and use several indicators that complement each other. It is suggested that the approach outlined here could be useful in developing a method for establishing a national chemical footprint

  19. Source of vacuum electromagnetic zero-point energy and Dirac's large numbers hypothesis

    International Nuclear Information System (INIS)

    Simaciu, I.; Dumitrescu, G.

    1993-01-01

    The stochastic electrodynamics states that zero-point fluctuation of the vacuum (ZPF) is an electromagnetic zero-point radiation with spectral density ρ(ω)=ℎω 3 / 2π 2 C 3 . Protons, free electrons and atoms are sources for this radiation. Each of them absorbs and emits energy by interacting with ZPF. At equilibrium ZPF radiation is scattered by dipoles.Scattered radiation spectral density is ρ(ω,r) ρ(ω).c.σ(ω) / 4πr 2 . Radiation of dipole spectral density of Universe is ρ ∫ 0 R nρ(ω,r)4πr 2 dr. But if σ atom P e σ=σ T then ρ ρ(ω)σ T R.n. Moreover if ρ=ρ(ω) then σ T Rn = 1. With R = G M/c 2 and σ T ≅(e 2 /m e c 2 ) 2 ∝ r e 2 then σ T .Rn 1 is equivalent to R/r e = e 2 /Gm p m e i.e. the cosmological coincidence discussed in the context of Dirac's large-numbers hypothesis. (Author)

  20. The ESID Online Database network.

    Science.gov (United States)

    Guzman, D; Veit, D; Knerr, V; Kindle, G; Gathmann, B; Eades-Perner, A M; Grimbacher, B

    2007-03-01

    Primary immunodeficiencies (PIDs) belong to the group of rare diseases. The European Society for Immunodeficiencies (ESID), is establishing an innovative European patient and research database network for continuous long-term documentation of patients, in order to improve the diagnosis, classification, prognosis and therapy of PIDs. The ESID Online Database is a web-based system aimed at data storage, data entry, reporting and the import of pre-existing data sources in an enterprise business-to-business integration (B2B). The online database is based on Java 2 Enterprise System (J2EE) with high-standard security features, which comply with data protection laws and the demands of a modern research platform. The ESID Online Database is accessible via the official website (http://www.esid.org/). Supplementary data are available at Bioinformatics online.

  1. Sources and methods to reconstruct past masting patterns in European oak species.

    Science.gov (United States)

    Szabó, Péter

    2012-01-01

    The irregular occurrence of good seed years in forest trees is known in many parts of the world. Mast year frequency in the past few decades can be examined through field observational studies; however, masting patterns in the more distant past are equally important in gaining a better understanding of long-term forest ecology. Past masting patterns can be studied through the examination of historical written sources. These pose considerable challenges, because data in them were usually not recorded with the aim of providing information about masting. Several studies examined masting in the deeper past, however, authors hardly ever considered the methodological implications of using and combining various source types. This paper provides a critical overview of the types of archival written that are available for the reconstruction of past masting patterns for European oak species and proposes a method to unify and evaluate different types of data. Available sources cover approximately eight centuries and can be put into two basic categories: direct observations on the amount of acorns and references to sums of money received in exchange for access to acorns. Because archival sources are highly different in origin and quality, the optimal solution for creating databases for past masting data is a three-point scale: zero mast, moderate mast, good mast. When larger amounts of data are available in a unified three-point-scale database, they can be used to test hypotheses about past masting frequencies, the driving forces of masting or regional masting patterns.

  2. Point, surface and volumetric heat sources in the thermal modelling of selective laser melting

    Science.gov (United States)

    Yang, Yabin; Ayas, Can

    2017-10-01

    Selective laser melting (SLM) is a powder based additive manufacturing technique suitable for producing high precision metal parts. However, distortions and residual stresses within products arise during SLM because of the high temperature gradients created by the laser heating. Residual stresses limit the load resistance of the product and may even lead to fracture during the built process. It is therefore of paramount importance to predict the level of part distortion and residual stress as a function of SLM process parameters which requires a reliable thermal modelling of the SLM process. Consequently, a key question arises which is how to describe the laser source appropriately. Reasonable simplification of the laser representation is crucial for the computational efficiency of the thermal model of the SLM process. In this paper, first a semi-analytical thermal modelling approach is described. Subsequently, the laser heating is modelled using point, surface and volumetric sources, in order to compare the influence of different laser source geometries on the thermal history prediction of the thermal model. The present work provides guidelines on appropriate representation of the laser source in the thermal modelling of the SLM process.

  3. The 1.4-2.7 micron spectrum of the point source at the galactic center

    Science.gov (United States)

    Treffers, R. R.; Fink, U.; Larson, H. P.; Gautier, T. N., III

    1976-01-01

    The spectrum of the 2-micron point source at the galactic center is presented over the range from 1.4 to 2.7 microns. The two-level-transition CO band heads are seen near 2.3 microns, confirming that the radiation from this source is due to a cool supergiant star. The heliocentric radial velocity is found to be - 173 (+ or -90) km/s and is consistent with the star being in orbit about a dense galactic nucleus. No evidence is found for Brackett-gamma emission, and no interstellar absorption features are seen. Upper limits for the column densities of interstellar H2, CH4, CO, and NH3 are derived.

  4. SNaX: A Database of Supernova X-Ray Light Curves

    Energy Technology Data Exchange (ETDEWEB)

    Ross, Mathias; Dwarkadas, Vikram V., E-mail: Mathias_Ross@msn.com, E-mail: vikram@oddjob.uchicago.edu [Astronomy and Astrophysics, University of Chicago, 5640 S Ellis Avenue, ERC 569, Chicago, IL 60637 (United States)

    2017-06-01

    We present the Supernova X-ray Database (SNaX), a compilation of the X-ray data from young supernovae (SNe). The database includes the X-ray fluxes and luminosities of young SNe, from days to years after outburst. The original goal and intent of this study was to present a database of Type IIn SNe (SNe IIn), which we have accomplished. Our ongoing goal is to expand the database to include all SNe for which published data are available. The database interface allows one to search for SNe using various criteria, plot all or selected data points, and download both the data and the plot. The plotting facility allows for significant customization. There is also a facility for the user to submit data that can be directly incorporated into the database. We include an option to fit the decay of any given SN light curve with a power-law. The database includes a conversion of most data points to a common 0.3–8 keV band so that SN light curves may be directly compared with each other. A mailing list has been set up to disseminate information about the database. We outline the structure and function of the database, describe its various features, and outline the plans for future expansion.

  5. SNaX: A Database of Supernova X-Ray Light Curves

    International Nuclear Information System (INIS)

    Ross, Mathias; Dwarkadas, Vikram V.

    2017-01-01

    We present the Supernova X-ray Database (SNaX), a compilation of the X-ray data from young supernovae (SNe). The database includes the X-ray fluxes and luminosities of young SNe, from days to years after outburst. The original goal and intent of this study was to present a database of Type IIn SNe (SNe IIn), which we have accomplished. Our ongoing goal is to expand the database to include all SNe for which published data are available. The database interface allows one to search for SNe using various criteria, plot all or selected data points, and download both the data and the plot. The plotting facility allows for significant customization. There is also a facility for the user to submit data that can be directly incorporated into the database. We include an option to fit the decay of any given SN light curve with a power-law. The database includes a conversion of most data points to a common 0.3–8 keV band so that SN light curves may be directly compared with each other. A mailing list has been set up to disseminate information about the database. We outline the structure and function of the database, describe its various features, and outline the plans for future expansion.

  6. Nutrient Losses from Non-Point Sources or from Unidentified Point Sources? Application Examples of the Smartphone Based Nitrate App.

    Science.gov (United States)

    Rozemeijer, J.; Ekkelenkamp, R.; van der Zaan, B.

    2017-12-01

    In 2016 Deltares launched the free to use Nitrate App which accurately reads and interprets nitrate test strips. The app directly displays the measured concentration and gives the option to share the result. Shared results are visualised in map functionality within the app and online. Since its introduction we've been seeing an increasing number of nitrate app applications. In this presentation we show some unanticipated types of application. The Nitrate App was originally intended to enable farmers to measure nitrate concentrations on their own farms. This may encourage farmers to talk to specialists about the right nutrient best management practices (BMP's) for their farm. Several groups of farmers have recently started to apply the Nitrate App and to discuss their results with each other and with the authorities. Nitrate concentration routings in catchments have proven to be another useful application. Within a day a person can generate a catchment scale nitrate concentration map identifying nitrate loss hotspots. In several routings in agricultural catchments clear point sources were found, for example at small scale manure processing plants. These routings proved that the Nitrate App can help water managers to target conservation practices more accurately to areas with the highest nitrate concentrations and loads. Other current applications are the screening of domestic water wells in California, the collection of extra measurements (also pH and NH4) in the National Monitoring Network for the Evaluation of the Manure Policy in the Netherlands, and several educational initiatives in cooperation with schools and universities.

  7. Experiment Databases

    Science.gov (United States)

    Vanschoren, Joaquin; Blockeel, Hendrik

    Next to running machine learning algorithms based on inductive queries, much can be learned by immediately querying the combined results of many prior studies. Indeed, all around the globe, thousands of machine learning experiments are being executed on a daily basis, generating a constant stream of empirical information on machine learning techniques. While the information contained in these experiments might have many uses beyond their original intent, results are typically described very concisely in papers and discarded afterwards. If we properly store and organize these results in central databases, they can be immediately reused for further analysis, thus boosting future research. In this chapter, we propose the use of experiment databases: databases designed to collect all the necessary details of these experiments, and to intelligently organize them in online repositories to enable fast and thorough analysis of a myriad of collected results. They constitute an additional, queriable source of empirical meta-data based on principled descriptions of algorithm executions, without reimplementing the algorithms in an inductive database. As such, they engender a very dynamic, collaborative approach to experimentation, in which experiments can be freely shared, linked together, and immediately reused by researchers all over the world. They can be set up for personal use, to share results within a lab or to create open, community-wide repositories. Here, we provide a high-level overview of their design, and use an existing experiment database to answer various interesting research questions about machine learning algorithms and to verify a number of recent studies.

  8. Brain Tumor Database, a free relational database for collection and analysis of brain tumor patient information.

    Science.gov (United States)

    Bergamino, Maurizio; Hamilton, David J; Castelletti, Lara; Barletta, Laura; Castellan, Lucio

    2015-03-01

    In this study, we describe the development and utilization of a relational database designed to manage the clinical and radiological data of patients with brain tumors. The Brain Tumor Database was implemented using MySQL v.5.0, while the graphical user interface was created using PHP and HTML, thus making it easily accessible through a web browser. This web-based approach allows for multiple institutions to potentially access the database. The BT Database can record brain tumor patient information (e.g. clinical features, anatomical attributes, and radiological characteristics) and be used for clinical and research purposes. Analytic tools to automatically generate statistics and different plots are provided. The BT Database is a free and powerful user-friendly tool with a wide range of possible clinical and research applications in neurology and neurosurgery. The BT Database graphical user interface source code and manual are freely available at http://tumorsdatabase.altervista.org. © The Author(s) 2013.

  9. Point X-ray sources in the SNR G 315.4-2.30 (MSH 14-63, RCW 86)

    Science.gov (United States)

    Gvaramadze, V. V.; Vikhlinin, A. A.

    2003-04-01

    We report the results of a search for a point X-ray source (stellar remnant) in the southwest protrusion of the supernova remnant G 315.4-2.30 (MSH 14-63, RCW 86) using the archival data of the Chandra X-ray Observatory. The search was motivated by a hypothesis that G 315.4-2.30 is the result of an off-centered cavity supernova explosion of a moving massive star, which ended its evolution just near the edge of the main-sequence wind-driven bubble. This hypothesis implies that the southwest protrusion in G 315.4-2.30 is the remainder of a pre-existing bow shock-like structure created by the interaction of the supernova progenitor's wind with the interstellar medium and that the actual location of the supernova blast center is near the center of this hemispherical structure. We have discovered two point X-ray sources in the ``proper" place. One of the sources has an optical counterpart with the photographic magnitude 13.38+/-0.40, while the spectrum of the source can be fitted with an optically thin plasma model. We interpret this source as a foreground active star of late spectral type. The second source has no optical counterpart to a limiting magnitude ~ 21. The spectrum of this source can be fitted almost equally well with several simple models (power law: photon index =1.87; two-temperature blackbody: kT1 =0.11 keV, R1 =2.34 km and kT2 =0.71 keV, R2 =0.06 km; blackbody plus power law: kT =0.07 keV, photon index =2.3). We interpret this source as a candidate stellar remnant (neutron star), while the photon index and non-thermal luminosity of the source (almost the same as those of the Vela pulsar and the recently discovered pulsar PSR J 0205+6449 in the supernova remnant 3C 58) suggest that it can be a young ``ordinary" pulsar.

  10. Iterative image reconstruction for positron emission tomography based on a detector response function estimated from point source measurements

    International Nuclear Information System (INIS)

    Tohme, Michel S; Qi Jinyi

    2009-01-01

    The accuracy of the system model in an iterative reconstruction algorithm greatly affects the quality of reconstructed positron emission tomography (PET) images. For efficient computation in reconstruction, the system model in PET can be factored into a product of a geometric projection matrix and sinogram blurring matrix, where the former is often computed based on analytical calculation, and the latter is estimated using Monte Carlo simulations. Direct measurement of a sinogram blurring matrix is difficult in practice because of the requirement of a collimated source. In this work, we propose a method to estimate the 2D blurring kernels from uncollimated point source measurements. Since the resulting sinogram blurring matrix stems from actual measurements, it can take into account the physical effects in the photon detection process that are difficult or impossible to model in a Monte Carlo (MC) simulation, and hence provide a more accurate system model. Another advantage of the proposed method over MC simulation is that it can easily be applied to data that have undergone a transformation to reduce the data size (e.g., Fourier rebinning). Point source measurements were acquired with high count statistics in a relatively fine grid inside the microPET II scanner using a high-precision 2D motion stage. A monotonically convergent iterative algorithm has been derived to estimate the detector blurring matrix from the point source measurements. The algorithm takes advantage of the rotational symmetry of the PET scanner and explicitly models the detector block structure. The resulting sinogram blurring matrix is incorporated into a maximum a posteriori (MAP) image reconstruction algorithm. The proposed method has been validated using a 3 x 3 line phantom, an ultra-micro resolution phantom and a 22 Na point source superimposed on a warm background. The results of the proposed method show improvements in both resolution and contrast ratio when compared with the MAP

  11. Kernel integration scatter model for parallel beam gamma camera and SPECT point source response

    International Nuclear Information System (INIS)

    Marinkovic, P.M.

    2001-01-01

    Scatter correction is a prerequisite for quantitative single photon emission computed tomography (SPECT). In this paper a kernel integration scatter Scatter correction is a prerequisite for quantitative SPECT. In this paper a kernel integration scatter model for parallel beam gamma camera and SPECT point source response based on Klein-Nishina formula is proposed. This method models primary photon distribution as well as first Compton scattering. It also includes a correction for multiple scattering by applying a point isotropic single medium buildup factor for the path segment between the point of scatter an the point of detection. Gamma ray attenuation in the object of imaging, based on known μ-map distribution, is considered too. Intrinsic spatial resolution of the camera is approximated by a simple Gaussian function. Collimator is modeled simply using acceptance angles derived from the physical dimensions of the collimator. Any gamma rays satisfying this angle were passed through the collimator to the crystal. Septal penetration and scatter in the collimator were not included in the model. The method was validated by comparison with Monte Carlo MCNP-4a numerical phantom simulation and excellent results were obtained. The physical phantom experiments, to confirm this method, are planed to be done. (author)

  12. DBGC: A Database of Human Gastric Cancer

    Science.gov (United States)

    Wang, Chao; Zhang, Jun; Cai, Mingdeng; Zhu, Zhenggang; Gu, Wenjie; Yu, Yingyan; Zhang, Xiaoyan

    2015-01-01

    The Database of Human Gastric Cancer (DBGC) is a comprehensive database that integrates various human gastric cancer-related data resources. Human gastric cancer-related transcriptomics projects, proteomics projects, mutations, biomarkers and drug-sensitive genes from different sources were collected and unified in this database. Moreover, epidemiological statistics of gastric cancer patients in China and clinicopathological information annotated with gastric cancer cases were also integrated into the DBGC. We believe that this database will greatly facilitate research regarding human gastric cancer in many fields. DBGC is freely available at http://bminfor.tongji.edu.cn/dbgc/index.do PMID:26566288

  13. Search for atmospheric muon-neutrinos and extraterrestric neutrino point sources in the 1997 AMANDA-B10 data

    International Nuclear Information System (INIS)

    Biron von Curland, A.

    2002-07-01

    The young field of high energy neutrino astronomy can be motivated by the search for the origin of the charged cosmic rays. Large astrophysical objects like AGNs or supernova remnants are candidates to accelerate hadrons which then can interact to eventually produce high energy neutrinos. Neutrino-induced muons can be detected via their emission of Cherenkov light in large neutrino telescopes like AMANDA. More than 10 9 atmospheric muon events and approximately 5000 atmospheric neutrino events were registered by AMANDA-B10 in 1997. Out of these, 223 atmospheric neutrino candidate events have been extracted. This data set contains approximately 15 background events. It allows to confirm the expected sensitivity of the detector towards neutrino events. A second set containing 369 (approximately 270 atmospheric neutrino events and 100 atmospheric muon events) was used to search for extraterrestrial neutrino point sources. Neither a binned search, nor a cluster search, nor a search for preselected sources gave indications for the existence of a strong neutrino point source. Based on this result, flux limits were derived. Assuming E ν -2 spectra, typical flux limits for selected sources of the order of Φ μ limit ∝ 10 -14 cm -2 s -1 for muons and Φ ν limit ∝ 10 -7 cm -2 s -1 for neutrinos have been obtained. (orig.)

  14. Enhanced Fire Events Database to Support Fire PRA

    International Nuclear Information System (INIS)

    Baranowsky, Patrick; Canavan, Ken; St. Germain, Shawn

    2010-01-01

    This paper provides a description of the updated and enhanced Fire Events Data Base (FEDB) developed by the Electric Power Research Institute (EPRI) in cooperation with the U.S. Nuclear Regulatory Commission (NRC). The FEDB is the principal source of fire incident operational data for use in fire PRAs. It provides a comprehensive and consolidated source of fire incident information for nuclear power plants operating in the U.S. The database classification scheme identifies important attributes of fire incidents to characterize their nature, causal factors, and severity consistent with available data. The database provides sufficient detail to delineate important plant specific attributes of the incidents to the extent practical. A significant enhancement to the updated FEDB is the reorganization and refinement of the database structure and data fields and fire characterization details added to more rigorously capture the nature and magnitude of the fire and damage to the ignition source and nearby equipment and structures.

  15. Revolution of open source and film making towards open film making

    OpenAIRE

    Löker, Koray

    2008-01-01

    Ankara : The Department of Communication and Design and the Institute of Economics and Social Sciences of Bilkent University, 2008. Thesis (Master's) -- Bilkent University, 2008. Includes bibliographical references leaves 74-80. This thesis is a critical analysis of self-proclaimed open source movie projects, Elephants Dream and The Digital Tipping Point. The theoretical framework derived from the new media discourse on film making, mainly based on Lev Manovich's database na...

  16. The CATDAT damaging earthquakes database

    Science.gov (United States)

    Daniell, J. E.; Khazai, B.; Wenzel, F.; Vervaeck, A.

    2011-08-01

    The global CATDAT damaging earthquakes and secondary effects (tsunami, fire, landslides, liquefaction and fault rupture) database was developed to validate, remove discrepancies, and expand greatly upon existing global databases; and to better understand the trends in vulnerability, exposure, and possible future impacts of such historic earthquakes. Lack of consistency and errors in other earthquake loss databases frequently cited and used in analyses was a major shortcoming in the view of the authors which needed to be improved upon. Over 17 000 sources of information have been utilised, primarily in the last few years, to present data from over 12 200 damaging earthquakes historically, with over 7000 earthquakes since 1900 examined and validated before insertion into the database. Each validated earthquake includes seismological information, building damage, ranges of social losses to account for varying sources (deaths, injuries, homeless, and affected), and economic losses (direct, indirect, aid, and insured). Globally, a slightly increasing trend in economic damage due to earthquakes is not consistent with the greatly increasing exposure. The 1923 Great Kanto (214 billion USD damage; 2011 HNDECI-adjusted dollars) compared to the 2011 Tohoku (>300 billion USD at time of writing), 2008 Sichuan and 1995 Kobe earthquakes show the increasing concern for economic loss in urban areas as the trend should be expected to increase. Many economic and social loss values not reported in existing databases have been collected. Historical GDP (Gross Domestic Product), exchange rate, wage information, population, HDI (Human Development Index), and insurance information have been collected globally to form comparisons. This catalogue is the largest known cross-checked global historic damaging earthquake database and should have far-reaching consequences for earthquake loss estimation, socio-economic analysis, and the global reinsurance field.

  17. The CATDAT damaging earthquakes database

    Directory of Open Access Journals (Sweden)

    J. E. Daniell

    2011-08-01

    Full Text Available The global CATDAT damaging earthquakes and secondary effects (tsunami, fire, landslides, liquefaction and fault rupture database was developed to validate, remove discrepancies, and expand greatly upon existing global databases; and to better understand the trends in vulnerability, exposure, and possible future impacts of such historic earthquakes.

    Lack of consistency and errors in other earthquake loss databases frequently cited and used in analyses was a major shortcoming in the view of the authors which needed to be improved upon.

    Over 17 000 sources of information have been utilised, primarily in the last few years, to present data from over 12 200 damaging earthquakes historically, with over 7000 earthquakes since 1900 examined and validated before insertion into the database. Each validated earthquake includes seismological information, building damage, ranges of social losses to account for varying sources (deaths, injuries, homeless, and affected, and economic losses (direct, indirect, aid, and insured.

    Globally, a slightly increasing trend in economic damage due to earthquakes is not consistent with the greatly increasing exposure. The 1923 Great Kanto ($214 billion USD damage; 2011 HNDECI-adjusted dollars compared to the 2011 Tohoku (>$300 billion USD at time of writing, 2008 Sichuan and 1995 Kobe earthquakes show the increasing concern for economic loss in urban areas as the trend should be expected to increase. Many economic and social loss values not reported in existing databases have been collected. Historical GDP (Gross Domestic Product, exchange rate, wage information, population, HDI (Human Development Index, and insurance information have been collected globally to form comparisons.

    This catalogue is the largest known cross-checked global historic damaging earthquake database and should have far-reaching consequences for earthquake loss estimation, socio-economic analysis, and the global

  18. Stationary sources of airborne lead: a comparison of emissions data for southern California.

    Science.gov (United States)

    Harris, Allison R; Fifarek, Brian J; Davidson, Cliff I; Blackmon, Rebecca Lankey

    2006-04-01

    Estimates for the air releases of lead from stationary point sources are considered for the South Coast Air Basin of California. We have examined four databases published by U.S. Environmental Protection Agency, the California Air Resources Board, and the South Coast Air Quality Management District. Our analysis indicates that none of the databases includes every emitting facility in the South Coast Air Basin of California and that other discrepancies among the databases exist. Additionally, the data have been analyzed for temporal variation, and some of the California Air Resources Board data are not current. The South Coast Air Quality Management District inventory covers 12 times more facilities in 2001 than in 1996. From this analysis, we conclude that all four of the databases would benefit by sharing data, increasing transparency, analyzing uncertainty, and standardizing emission estimation methods.

  19. HATCHES - a thermodynamic database and management system

    International Nuclear Information System (INIS)

    Cross, J.E.; Ewart, F.T.

    1990-03-01

    The Nirex Safety Assessment Research Programme has been compiling the thermodynamic data necessary to allow simulations of the aqueous behaviour of the elements important to radioactive waste disposal to be made. These data have been obtained from the literature, when available, and validated for the conditions of interest by experiment. In order to maintain these data in an accessible form and to satisfy quality assurance on all data used for assessments, a database has been constructed which resides on a personal computer operating under MS-DOS using the Ashton-Tate dBase III program. This database contains all the input data fields required by the PHREEQE program and, in addition, a body of text which describes the source of the data and the derivation of the PHREEQE input parameters from the source data. The HATCHES system consists of this database, a suite of programs to facilitate the searching and listing of data and a further suite of programs to convert the dBase III files to PHREEQE database format. (Author)

  20. Calculation of total counting efficiency of a NaI(Tl) detector by hybrid Monte-Carlo method for point and disk sources

    Energy Technology Data Exchange (ETDEWEB)

    Yalcin, S. [Education Faculty, Kastamonu University, 37200 Kastamonu (Turkey)], E-mail: yalcin@gazi.edu.tr; Gurler, O.; Kaynak, G. [Department of Physics, Faculty of Arts and Sciences, Uludag University, Gorukle Campus, 16059 Bursa (Turkey); Gundogdu, O. [Department of Physics, School of Engineering and Physical Sciences, University of Surrey, Guildford GU2 7XH (United Kingdom)

    2007-10-15

    This paper presents results on the total gamma counting efficiency of a NaI(Tl) detector from point and disk sources. The directions of photons emitted from the source were determined by Monte-Carlo techniques and the photon path lengths in the detector were determined by analytic equations depending on photon directions. This is called the hybrid Monte-Carlo method where analytical expressions are incorporated into the Monte-Carlo simulations. A major advantage of this technique is the short computation time compared to other techniques on similar computational platforms. Another advantage is the flexibility for inputting detector-related parameters (such as source-detector distance, detector radius, source radius, detector linear attenuation coefficient) into the algorithm developed, thus making it an easy and flexible method to apply to other detector systems and configurations. The results of the total counting efficiency model put forward for point and disc sources were compared with the previous work reported in the literature.

  1. Calculation of total counting efficiency of a NaI(Tl) detector by hybrid Monte-Carlo method for point and disk sources

    International Nuclear Information System (INIS)

    Yalcin, S.; Gurler, O.; Kaynak, G.; Gundogdu, O.

    2007-01-01

    This paper presents results on the total gamma counting efficiency of a NaI(Tl) detector from point and disk sources. The directions of photons emitted from the source were determined by Monte-Carlo techniques and the photon path lengths in the detector were determined by analytic equations depending on photon directions. This is called the hybrid Monte-Carlo method where analytical expressions are incorporated into the Monte-Carlo simulations. A major advantage of this technique is the short computation time compared to other techniques on similar computational platforms. Another advantage is the flexibility for inputting detector-related parameters (such as source-detector distance, detector radius, source radius, detector linear attenuation coefficient) into the algorithm developed, thus making it an easy and flexible method to apply to other detector systems and configurations. The results of the total counting efficiency model put forward for point and disc sources were compared with the previous work reported in the literature

  2. The CERN accelerator measurement database: on the road to federation

    International Nuclear Information System (INIS)

    Roderick, C.; Billen, R.; Gourber-Pace, M.; Hoibian, N.; Peryt, M.

    2012-01-01

    The Measurement database, acting as short-term central persistence and front-end of the CERN accelerator Logging Service, receives billions of time-series data per day for 200000+ signals. A variety of data acquisition systems on hundreds of front-end computers publish source data that eventually end up being logged in the Measurement database. As part of a federated approach to data management, information about source devices are defined in a Configuration database, whilst the signals to be logged are defined in the Measurement database. A mapping, which is often complex and subject to change/extension, is required in order to subscribe to the source devices, and write the published data to the corresponding named signals. Since 2005, this mapping was done by means of dozens of XML files, which were manually maintained by multiple persons, resulting in a configuration that was error prone. In 2010 this configuration was fully centralized in the Measurement database itself, reducing significantly the complexity and the actors in the process. Furthermore, logging processes immediately pick up modified configurations via JMS based notifications sent directly from the database. This paper will describe the architecture and the benefits of current implementation, as well as the next steps on the road to a fully federated solution. (authors)

  3. A KINETIC DATABASE FOR ASTROCHEMISTRY (KIDA)

    International Nuclear Information System (INIS)

    Wakelam, V.; Pavone, B.; Hébrard, E.; Hersant, F.; Herbst, E.; Loison, J.-C.; Chandrasekaran, V.; Bergeat, A.; Smith, I. W. M.; Adams, N. G.; Bacchus-Montabonel, M.-C.; Béroff, K.; Bierbaum, V. M.; Chabot, M.; Dalgarno, A.; Van Dishoeck, E. F.; Faure, A.; Geppert, W. D.; Gerlich, D.; Galli, D.

    2012-01-01

    We present a novel chemical database for gas-phase astrochemistry. Named the KInetic Database for Astrochemistry (KIDA), this database consists of gas-phase reactions with rate coefficients and uncertainties that will be vetted to the greatest extent possible. Submissions of measured and calculated rate coefficients are welcome, and will be studied by experts before inclusion into the database. Besides providing kinetic information for the interstellar medium, KIDA is planned to contain such data for planetary atmospheres and for circumstellar envelopes. Each year, a subset of the reactions in the database (kida.uva) will be provided as a network for the simulation of the chemistry of dense interstellar clouds with temperatures between 10 K and 300 K. We also provide a code, named Nahoon, to study the time-dependent gas-phase chemistry of zero-dimensional and one-dimensional interstellar sources.

  4. XML databases and the semantic web

    CERN Document Server

    Thuraisingham, Bhavani

    2002-01-01

    Efficient access to data, sharing data, extracting information from data, and making use of the information have become urgent needs for today''s corporations. With so much data on the Web, managing it with conventional tools is becoming almost impossible. New tools and techniques are necessary to provide interoperability as well as warehousing between multiple data sources and systems, and to extract information from the databases. XML Databases and the Semantic Web focuses on critical and new Web technologies needed for organizations to carry out transactions on the Web, to understand how to use the Web effectively, and to exchange complex documents on the Web.This reference for database administrators, database designers, and Web designers working in tandem with database technologists covers three emerging technologies of significant impact for electronic business: Extensible Markup Language (XML), semi-structured databases, and the semantic Web. The first two parts of the book explore these emerging techn...

  5. Dynamic Control of Particle Deposition in Evaporating Droplets by an External Point Source of Vapor.

    Science.gov (United States)

    Malinowski, Robert; Volpe, Giovanni; Parkin, Ivan P; Volpe, Giorgio

    2018-02-01

    The deposition of particles on a surface by an evaporating sessile droplet is important for phenomena as diverse as printing, thin-film deposition, and self-assembly. The shape of the final deposit depends on the flows within the droplet during evaporation. These flows are typically determined at the onset of the process by the intrinsic physical, chemical, and geometrical properties of the droplet and its environment. Here, we demonstrate deterministic emergence and real-time control of Marangoni flows within the evaporating droplet by an external point source of vapor. By varying the source location, we can modulate these flows in space and time to pattern colloids on surfaces in a controllable manner.

  6. Trends in Solar energy Driven Vertical Ground Source Heat Pump Systems in Sweden - An Analysis Based on the Swedish Well Database

    Science.gov (United States)

    Juhlin, K.; Gehlin, S.

    2016-12-01

    Sweden is a world leader in developing and using vertical ground source heat pump (GSHP) technology. GSHP systems extract passively stored solar energy in the ground and the Earth's natural geothermal energy. Geothermal energy is an admitted renewable energy source in Sweden since 2007 and is the third largest renewable energy source in the country today. The Geological Survey of Sweden (SGU) is the authority in Sweden that provides open access geological data of rock, soil and groundwater for the public. All wells drilled must be registered in the SGU Well Database and it is the well driller's duty to submit registration of drilled wells.Both active and passive geothermal energy systems are in use. Large GSHP systems, with at least 20 boreholes, are active geothermal energy systems. Energy is stored in the ground which allows both comfort heating and cooling to be extracted. Active systems are therefore relevant for larger properties and industrial buildings. Since 1978 more than 600 000 wells (water wells, GSHP boreholes etc) have been registered in the Well Database, with around 20 000 new registrations per year. Of these wells an estimated 320 000 wells are registered as GSHP boreholes. The vast majority of these boreholes are single boreholes for single-family houses. The number of properties with registered vertical borehole GSHP installations amounts to approximately 243 000. Of these sites between 300-350 are large GSHP systems with at least 20 boreholes. While the increase in number of new registrations for smaller homes and households has slowed down after the rapid development in the 80's and 90's, the larger installations for commercial and industrial buildings have increased in numbers over the last ten years. This poster uses data from the SGU Well Database to quantify and analyze the trends in vertical GSHP systems reported between 1978-2015 in Sweden, with special focus on large systems. From the new aggregated data, conclusions can be drawn about

  7. Issues in Big-Data Database Systems

    Science.gov (United States)

    2014-06-01

    that big data will not be manageable using conventional relational database technology, and it is true that alternative paradigms, such as NoSQL systems...conventional relational database technology, and it is true that alternative paradigms, such as NoSQL systems and search engines, have much to offer...scale well, and because integration with external data sources is so difficult. NoSQL systems are more open to this integration, and provide excellent

  8. CAE meteorological database for the PC CREAM program. Atmospheric dilution factor in different points of the CAE (Centro Atomico Ezeiza) and of the argentine nuclear power plants

    International Nuclear Information System (INIS)

    Amado, Valeria A.

    2007-01-01

    In the first part of this work, the EZEIZA.MET file, with the meteorological database of the surroundings of the Ezeiza Atomic Center, is prepared and incorporated into the library of the PC CREAM program. This program was developed by the National Radiological Protection Board and the European Union. Information provided by the National Meteorological Service was used, corresponding to the Ezeiza Meteorological Station during the period 1996-2005. In the second part, a methodology to estimate the atmospheric dilution factor at a point using the PLUME module of the PC CREAM, is presented. The developed methodology was used to estimate the dilution factor at points close to the Ezeiza Atomic Center and nuclear power plants Atucha I and Embalse. The developed methodology was used to estimate the dilution factor at points close to the Ezeiza Atomic Center and nuclear power plants Atucha I and Embalse. In the first case the file with the generated meteorological database is used, whereas for the nuclear power plants the already existing ATUCHALO.MET and EMBALSE.MET files are used. The dilution factors obtained are compared with those obtained in previous work. The proposed methodology is a useful tool to estimate the dilution factors in a simple and systematic way, and simultaneously allows the update of the meteorological information used in the estimations. (author) [es

  9. A systematic analysis of the Braitenberg vehicle 2b for point-like stimulus sources

    International Nuclear Information System (INIS)

    Rañó, Iñaki

    2012-01-01

    Braitenberg vehicles have been used experimentally for decades in robotics with limited empirical understanding. This paper presents the first mathematical model of the vehicle 2b, displaying so-called aggression behaviour, and analyses the possible trajectories for point-like smooth stimulus sources. This sensory-motor steering control mechanism is used to implement biologically grounded target approach, target-seeking or obstacle-avoidance behaviour. However, the analysis of the resulting model reveals that complex and unexpected trajectories can result even for point-like stimuli. We also prove how the implementation of the controller and the vehicle morphology interact to affect the behaviour of the vehicle. This work provides a better understanding of Braitenberg vehicle 2b, explains experimental results and paves the way for a formally grounded application on robotics as well as for a new way of understanding target seeking in biology. (paper)

  10. Preliminary limits on the flux of muon neutrinos from extraterrestrial point sources

    International Nuclear Information System (INIS)

    Bionta, R.M.; Blewitt, G.; Bratton, C.B.

    1985-01-01

    We present the arrival directions of 117 upward-going muon events collected with the IMB proton lifetime detector during 317 days of live detector operation. The rate of upward-going muons observed in our detector was found to be consistent with the rate expected from atmospheric neutrino production. The upper limit on the total flux of extraterrestrial neutrinos >1 GeV is 2 -sec. Using our data and a Monte Carlo simulation of high energy muon production in the earth surrounding the detector, we place limits on the flux of neutrinos from a point source in the Vela X-2 system of 2 -sec with E > 1 GeV. 6 refs., 5 figs

  11. Considering a point-source in a regional air pollution model; Prise en compte d`une source ponctuelle dans un modele regional de pollution atmospherique

    Energy Technology Data Exchange (ETDEWEB)

    Lipphardt, M.

    1997-06-19

    This thesis deals with the development and validation of a point-source plume model, with the aim to refine the representation of intensive point-source emissions in regional-scale air quality models. The plume is modelled at four levels of increasing complexity, from a modified Gaussian plume model to the Freiberg and Lusis ring model. Plume elevation is determined by Netterville`s plume rise model, using turbulence and atmospheric stability parameters. A model for the effect of a fine-scale turbulence on the mean concentrations in the plume is developed and integrated in the ring model. A comparison between results with and without considering micro-mixing shows the importance of this effect in a chemically reactive plume. The plume model is integrated into the Eulerian transport/chemistry model AIRQUAL, using an interface between Airqual and the sub-model, and interactions between the two scales are described. A simulation of an air pollution episode over Paris is carried out, showing that the utilization of such a sub-scale model improves the accuracy of the air quality model

  12. A new three-dimensional nonscanning laser imaging system based on the illumination pattern of a point-light-source array

    Science.gov (United States)

    Xia, Wenze; Ma, Yayun; Han, Shaokun; Wang, Yulin; Liu, Fei; Zhai, Yu

    2018-06-01

    One of the most important goals of research on three-dimensional nonscanning laser imaging systems is the improvement of the illumination system. In this paper, a new three-dimensional nonscanning laser imaging system based on the illumination pattern of a point-light-source array is proposed. This array is obtained using a fiber array connected to a laser array with each unit laser having independent control circuits. This system uses a point-to-point imaging process, which is realized using the exact corresponding optical relationship between the point-light-source array and a linear-mode avalanche photodiode array detector. The complete working process of this system is explained in detail, and the mathematical model of this system containing four equations is established. A simulated contrast experiment and two real contrast experiments which use the simplified setup without a laser array are performed. The final results demonstrate that unlike a conventional three-dimensional nonscanning laser imaging system, the proposed system meets all the requirements of an eligible illumination system. Finally, the imaging performance of this system is analyzed under defocusing situations, and analytical results show that the system has good defocusing robustness and can be easily adjusted in real applications.

  13. An open source web interface for linking models to infrastructure system databases

    Science.gov (United States)

    Knox, S.; Mohamed, K.; Harou, J. J.; Rheinheimer, D. E.; Medellin-Azuara, J.; Meier, P.; Tilmant, A.; Rosenberg, D. E.

    2016-12-01

    Models of networked engineered resource systems such as water or energy systems are often built collaboratively with developers from different domains working at different locations. These models can be linked to large scale real world databases, and they are constantly being improved and extended. As the development and application of these models becomes more sophisticated, and the computing power required for simulations and/or optimisations increases, so has the need for online services and tools which enable the efficient development and deployment of these models. Hydra Platform is an open source, web-based data management system, which allows modellers of network-based models to remotely store network topology and associated data in a generalised manner, allowing it to serve multiple disciplines. Hydra Platform uses a web API using JSON to allow external programs (referred to as `Apps') to interact with its stored networks and perform actions such as importing data, running models, or exporting the networks to different formats. Hydra Platform supports multiple users accessing the same network and has a suite of functions for managing users and data. We present ongoing development in Hydra Platform, the Hydra Web User Interface, through which users can collaboratively manage network data and models in a web browser. The web interface allows multiple users to graphically access, edit and share their networks, run apps and view results. Through apps, which are located on the server, the web interface can give users access to external data sources and models without the need to install or configure any software. This also ensures model results can be reproduced by removing platform or version dependence. Managing data and deploying models via the web interface provides a way for multiple modellers to collaboratively manage data, deploy and monitor model runs and analyse results.

  14. Development of a Consumer Product Ingredient Database for ...

    Science.gov (United States)

    Consumer products are a primary source of chemical exposures, yet little structured information is available on the chemical ingredients of these products and the concentrations at which ingredients are present. To address this data gap, we created a database of chemicals in consumer products using product Material Safety Data Sheets (MSDSs) publicly provided by a large retailer. The resulting database represents 1797 unique chemicals mapped to 8921 consumer products and a hierarchy of 353 consumer product “use categories” within a total of 15 top-level categories. We examine the utility of this database and discuss ways in which it will support (i) exposure screening and prioritization, (ii) generic or framework formulations for several indoor/consumer product exposure modeling initiatives, (iii) candidate chemical selection for monitoring near field exposure from proximal sources, and (iv) as activity tracers or ubiquitous exposure sources using “chemical space” map analyses. Chemicals present at high concentrations and across multiple consumer products and use categories that hold high exposure potential are identified. Our database is publicly available to serve regulators, retailers, manufacturers, and the public for predictive screening of chemicals in new and existing consumer products on the basis of exposure and risk. The National Exposure Research Laboratory’s (NERL’s) Human Exposure and Atmospheric Sciences Division (HEASD) conducts resear

  15. Prevention and Control of Agricultural Non-Point Source Pollutions in UK and Suggestions to China

    OpenAIRE

    Liu, Kun; Ren, Tianzhi; Wu, Wenliang; Meng, Fanquiao; Bellarby, Jessica; Smith, Laurence

    2016-01-01

    Currently, the world is facing challenges of maintaining food production growth while improving agricultural ecological environmental quality. The prevention and control of agricultural non-point source pollution, a key component of these challenges, is a systematic program which integrates many factors such as technology and its extension, relevant regulation and policies. In the project of UK-China Sustainable Agriculture Innovation Network, we undertook a comprehensive analysis of the prev...

  16. Aurorasaurus Database of Real-Time, Soft-Sensor Sourced Aurora Data for Space Weather Research

    Science.gov (United States)

    Kosar, B.; MacDonald, E.; Heavner, M.

    2017-12-01

    Aurorasaurus is an innovative citizen science project focused on two fundamental objectives i.e., collecting real-time, ground-based signals of auroral visibility from citizen scientists (soft-sensors) and incorporating this new type of data into scientific investigations pertaining to aurora. The project has been live since the Fall of 2014, and as of Summer 2017, the database compiled approximately 12,000 observations (5295 direct reports and 6413 verified tweets). In this presentation, we will focus on demonstrating the utility of this robust science quality data for space weather research needs. These data scale with the size of the event and are well-suited to capture the largest, rarest events. Emerging state-of-the-art computational methods based on statistical inference such as machine learning frameworks and data-model integration methods can offer new insights that could potentially lead to better real-time assessment and space weather prediction when citizen science data are combined with traditional sources.

  17. A Generative Approach for Building Database Federations

    Directory of Open Access Journals (Sweden)

    Uwe Hohenstein

    1999-11-01

    Full Text Available A comprehensive, specification-based approach for building database federations is introduced that supports an integrated ODMG2.0 conforming access to heterogeneous data sources seamlessly done in C++. The approach is centered around several generators. A first set of generators produce ODMG adapters for local sources in order to homogenize them. Each adapter represents an ODMG view and supports the ODMG manipulation and querying. The adapters can be plugged into a federation framework. Another generator produces an homogeneous and uniform view by putting an ODMG conforming federation layer on top of the adapters. Input to these generators are schema specifications. Schemata are defined in corresponding specification languages. There are languages to homogenize relational and object-oriented databases, as well as ordinary file systems. Any specification defines an ODMG schema and relates it to an existing data source. An integration language is then used to integrate the schemata and to build system-spanning federated views thereupon. The generative nature provides flexibility with respect to schema modification of component databases. Any time a schema changes, only the specification has to be adopted; new adapters are generated automatically

  18. Performance summary on a high power dense plasma focus x-ray lithography point source producing 70 nm line features in AlGaAs microcircuits

    International Nuclear Information System (INIS)

    Petr, Rodney; Bykanov, Alexander; Freshman, Jay; Reilly, Dennis; Mangano, Joseph; Roche, Maureen; Dickenson, Jason; Burte, Mitchell; Heaton, John

    2004-01-01

    A high average power dense plasma focus (DPF), x-ray point source has been used to produce ∼70 nm line features in AlGaAs-based monolithic millimeter-wave integrated circuits (MMICs). The DPF source has produced up to 12 J per pulse of x-ray energy into 4π steradians at ∼1 keV effective wavelength in ∼2 Torr neon at pulse repetition rates up to 60 Hz, with an effective x-ray yield efficiency of ∼0.8%. Plasma temperature and electron concentration are estimated from the x-ray spectrum to be ∼170 eV and ∼5·10 19 cm -3 , respectively. The x-ray point source utilizes solid-state pulse power technology to extend the operating lifetime of electrodes and insulators in the DPF discharge. By eliminating current reversals in the DPF head, an anode electrode has demonstrated a lifetime of more than 5 million shots. The x-ray point source has also been operated continuously for 8 h run times at 27 Hz average pulse recurrent frequency. Measurements of shock waves produced by the plasma discharge indicate that overpressure pulses must be attenuated before a collimator can be integrated with the DPF point source

  19. KaBOB: ontology-based semantic integration of biomedical databases.

    Science.gov (United States)

    Livingston, Kevin M; Bada, Michael; Baumgartner, William A; Hunter, Lawrence E

    2015-04-23

    The ability to query many independent biological databases using a common ontology-based semantic model would facilitate deeper integration and more effective utilization of these diverse and rapidly growing resources. Despite ongoing work moving toward shared data formats and linked identifiers, significant problems persist in semantic data integration in order to establish shared identity and shared meaning across heterogeneous biomedical data sources. We present five processes for semantic data integration that, when applied collectively, solve seven key problems. These processes include making explicit the differences between biomedical concepts and database records, aggregating sets of identifiers denoting the same biomedical concepts across data sources, and using declaratively represented forward-chaining rules to take information that is variably represented in source databases and integrating it into a consistent biomedical representation. We demonstrate these processes and solutions by presenting KaBOB (the Knowledge Base Of Biomedicine), a knowledge base of semantically integrated data from 18 prominent biomedical databases using common representations grounded in Open Biomedical Ontologies. An instance of KaBOB with data about humans and seven major model organisms can be built using on the order of 500 million RDF triples. All source code for building KaBOB is available under an open-source license. KaBOB is an integrated knowledge base of biomedical data representationally based in prominent, actively maintained Open Biomedical Ontologies, thus enabling queries of the underlying data in terms of biomedical concepts (e.g., genes and gene products, interactions and processes) rather than features of source-specific data schemas or file formats. KaBOB resolves many of the issues that routinely plague biomedical researchers intending to work with data from multiple data sources and provides a platform for ongoing data integration and development and for

  20. Five-Level Z-Source Neutral Point-Clamped Inverter

    DEFF Research Database (Denmark)

    Gao, F.; Loh, P.C.; Blaabjerg, Frede

    2007-01-01

    This paper proposes a five-level Z-source neutralpoint- clamped (NPC) inverter with two Z-source networks functioning as intermediate energy storages coupled between dc sources and NPC inverter circuitry. Analyzing the operational principles of Z-source network with partial dclink shoot......-through scheme reveals the hidden theories in the five-level Z-source NPC inverter unlike the operational principle appeared in the general two-level Z-source inverter, so that the five-level Z-source NPC inverter can be designed with the modulation of carrier-based phase disposition (PD) or alternative phase...

  1. On-site meteorological instrumentation requirements to characterize diffusion from point sources: workshop report. Final report Sep 79-Sep 80

    International Nuclear Information System (INIS)

    Strimaitis, D.; Hoffnagle, G.; Bass, A.

    1981-04-01

    Results of a workshop entitled 'On-Site Meteorological Instrumentation Requirements to Characterize Diffusion from Point Sources' are summarized and reported. The workshop was sponsored by the U.S. Environmental Protection Agency in Raleigh, North Carolina, on January 15-17, 1980. Its purpose was to provide EPA with a thorough examination of the meteorological instrumentation and data collection requirements needed to characterize airborne dispersion of air contaminants from point sources and to recommend, based on an expert consensus, specific measurement technique and accuracies. Secondary purposes of the workshop were to (1) make recommendations to the National Weather Service (NWS) about collecting and archiving meteorological data that would best support air quality dispersion modeling objectives and (2) make recommendations on standardization of meteorological data reporting and quality assurance programs

  2. [Estimation of urban non-point source pollution loading and its factor analysis in the Pearl River Delta].

    Science.gov (United States)

    Liao, Yi-Shan; Zhuo, Mu-Ning; Li, Ding-Qiang; Guo, Tai-Long

    2013-08-01

    In the Pearl Delta region, urban rivers have been seriously polluted, and the input of non-point source pollution materials, such as chemical oxygen demand (COD), into rivers cannot be neglected. During 2009-2010, the water qualities at eight different catchments in the Fenjiang River of Foshan city were monitored, and the COD loads for eight rivulet sewages were calculated in respect of different rainfall conditions. Interesting results were concluded in our paper. The rainfall and landuse type played important roles in the COD loading, with greater influence of rainfall than landuse type. Consequently, a COD loading formula was constructed that was defined as a function of runoff and landuse type that were derived SCS model and land use map. Loading of COD could be evaluated and predicted with the constructed formula. The mean simulation accuracy for single rainfall event was 75.51%. Long-term simulation accuracy was better than that of single rainfall. In 2009, the estimated COD loading and its loading intensity were 8 053 t and 339 kg x (hm2 x a)(-1), and the industrial land was regarded as the main source of COD pollution area. The severe non-point source pollution such as COD in Fenjiang River must be paid more attention in the future.

  3. E-MSD: the European Bioinformatics Institute Macromolecular Structure Database.

    Science.gov (United States)

    Boutselakis, H; Dimitropoulos, D; Fillon, J; Golovin, A; Henrick, K; Hussain, A; Ionides, J; John, M; Keller, P A; Krissinel, E; McNeil, P; Naim, A; Newman, R; Oldfield, T; Pineda, J; Rachedi, A; Copeland, J; Sitnov, A; Sobhany, S; Suarez-Uruena, A; Swaminathan, J; Tagari, M; Tate, J; Tromm, S; Velankar, S; Vranken, W

    2003-01-01

    The E-MSD macromolecular structure relational database (http://www.ebi.ac.uk/msd) is designed to be a single access point for protein and nucleic acid structures and related information. The database is derived from Protein Data Bank (PDB) entries. Relational database technologies are used in a comprehensive cleaning procedure to ensure data uniformity across the whole archive. The search database contains an extensive set of derived properties, goodness-of-fit indicators, and links to other EBI databases including InterPro, GO, and SWISS-PROT, together with links to SCOP, CATH, PFAM and PROSITE. A generic search interface is available, coupled with a fast secondary structure domain search tool.

  4. Developments in diffraction databases

    International Nuclear Information System (INIS)

    Jenkins, R.

    1999-01-01

    Full text: There are a number of databases available to the diffraction community. Two of the more important of these are the Powder Diffraction File (PDF) maintained by the International Centre for Diffraction Data (ICDD), and the Inorganic Crystal Structure Database (ICSD) maintained by Fachsinformationzentrum (FIZ, Karlsruhe). In application, the PDF has been used as an indispensable tool in phase identification and identification of unknowns. The ICSD database has extensive and explicit reference to the structures of compounds: atomic coordinates, space group and even thermal vibration parameters. A similar database, but for organic compounds, is maintained by the Cambridge Crystallographic Data Centre. These databases are often used as independent sources of information. However, little thought has been given on how to exploit the combined properties of structural database tools. A recently completed agreement between ICDD and FIZ, plus ICDD and Cambridge, provides a first step in complementary use of the PDF and the ICSD databases. The focus of this paper (as indicated below) is to examine ways of exploiting the combined properties of both databases. In 1996, there were approximately 76,000 entries in the PDF and approximately 43,000 entries in the ICSD database. The ICSD database has now been used to calculate entries in the PDF. Thus, to derive d-spacing and peak intensity data requires the synthesis of full diffraction patterns, i.e., we use the structural data in the ICSD database and then add instrumental resolution information. The combined data from PDF and ICSD can be effectively used in many ways. For example, we can calculate PDF data for an ideally random crystal distribution and also in the absence of preferred orientation. Again, we can use systematic studies of intermediate members in solid solutions series to help produce reliable quantitative phase analyses. In some cases, we can study how solid solution properties vary with composition and

  5. Investigating the effects of point source and nonpoint source pollution on the water quality of the East River (Dongjiang) in South China

    Science.gov (United States)

    Wu, Yiping; Chen, Ji

    2013-01-01

    Understanding the physical processes of point source (PS) and nonpoint source (NPS) pollution is critical to evaluate river water quality and identify major pollutant sources in a watershed. In this study, we used the physically-based hydrological/water quality model, Soil and Water Assessment Tool, to investigate the influence of PS and NPS pollution on the water quality of the East River (Dongjiang in Chinese) in southern China. Our results indicate that NPS pollution was the dominant contribution (>94%) to nutrient loads except for mineral phosphorus (50%). A comprehensive Water Quality Index (WQI) computed using eight key water quality variables demonstrates that water quality is better upstream than downstream despite the higher level of ammonium nitrogen found in upstream waters. Also, the temporal (seasonal) and spatial distributions of nutrient loads clearly indicate the critical time period (from late dry season to early wet season) and pollution source areas within the basin (middle and downstream agricultural lands), which resource managers can use to accomplish substantial reduction of NPS pollutant loadings. Overall, this study helps our understanding of the relationship between human activities and pollutant loads and further contributes to decision support for local watershed managers to protect water quality in this region. In particular, the methods presented such as integrating WQI with watershed modeling and identifying the critical time period and pollutions source areas can be valuable for other researchers worldwide.

  6. Iterative closest normal point for 3D face recognition.

    Science.gov (United States)

    Mohammadzade, Hoda; Hatzinakos, Dimitrios

    2013-02-01

    The common approach for 3D face recognition is to register a probe face to each of the gallery faces and then calculate the sum of the distances between their points. This approach is computationally expensive and sensitive to facial expression variation. In this paper, we introduce the iterative closest normal point method for finding the corresponding points between a generic reference face and every input face. The proposed correspondence finding method samples a set of points for each face, denoted as the closest normal points. These points are effectively aligned across all faces, enabling effective application of discriminant analysis methods for 3D face recognition. As a result, the expression variation problem is addressed by minimizing the within-class variability of the face samples while maximizing the between-class variability. As an important conclusion, we show that the surface normal vectors of the face at the sampled points contain more discriminatory information than the coordinates of the points. We have performed comprehensive experiments on the Face Recognition Grand Challenge database, which is presently the largest available 3D face database. We have achieved verification rates of 99.6 and 99.2 percent at a false acceptance rate of 0.1 percent for the all versus all and ROC III experiments, respectively, which, to the best of our knowledge, have seven and four times less error rates, respectively, compared to the best existing methods on this database.

  7. Basic database performance tuning - developer's perspective

    CERN Document Server

    Kwiatek, Michal

    2008-01-01

    This lecture discusses selected database performance issues from the developer's point of view: connection overhead, bind variables and SQL injection, making most of the optimizer with up-to-date statistics, reading execution plans. Prior knowledge of SQL is expected.

  8. Allegheny County Cell Tower Points

    Data.gov (United States)

    Allegheny County / City of Pittsburgh / Western PA Regional Data Center — This dataset portrays cell tower locations as points in Allegheny County. The dataset is based on outbuilding codes in the Property Assessment Parcel Database used...

  9. Guidelines for the implementation of an open source information system

    Energy Technology Data Exchange (ETDEWEB)

    Doak, J.; Howell, J.A.

    1995-08-01

    This work was initially performed for the International Atomic Energy Agency (IAEA) to help with the Open Source Task of the 93 + 2 Initiative; however, the information should be of interest to anyone working with open sources. The authors cover all aspects of an open source information system (OSIS) including, for example, identifying relevant sources, understanding copyright issues, and making information available to analysts. They foresee this document as a reference point that implementors of a system could augment for their particular needs. The primary organization of this document focuses on specific aspects, or components, of an OSIS; they describe each component and often make specific recommendations for its implementation. This document also contains a section discussing the process of collecting open source data and a section containing miscellaneous information. The appendix contains a listing of various providers, producers, and databases that the authors have come across in their research.

  10. ARTI Refrigerant Database

    Energy Technology Data Exchange (ETDEWEB)

    Cain, J.M. (Calm (James M.), Great Falls, VA (United States))

    1993-04-30

    The Refrigerant Database consolidates and facilitates access to information to assist industry in developing equipment using alternative refrigerants. The underlying purpose is to accelerate phase out of chemical compounds of environmental concern. The database provides bibliographic citations and abstracts for publications that may be useful in research and design of air-conditioning and refrigeration equipment. The complete documents are not included. The database identifies sources of specific information on R-32, R-123, R-124, R-125, R-134, R-134a, R-141b, R-142b, R-143a, R-152a, R-245ca, R-290 (propane), R-717 (ammonia), ethers, and others as well as azeotropic and zeotropic blends of these fluids. It addresses lubricants including alkylbenzene, polyalkylene glycol, ester, and other synthetics as well as mineral oils. It also references documents addressing compatibility of refrigerants and lubricants with metals, plastics, elastomers, motor insulation, and other materials used in refrigerant circuits. Incomplete citations or abstracts are provided for some documents to accelerate availability of the information and will be completed or replaced in future updates.

  11. ARTI refrigerant database

    Energy Technology Data Exchange (ETDEWEB)

    Calm, J.M.

    1997-02-01

    The Refrigerant Database is an information system on alternative refrigerants, associated lubricants, and their use in air conditioning and refrigeration. It consolidates and facilitates access to property, compatibility, environmental, safety, application and other information. It provides corresponding information on older refrigerants, to assist manufacturers and those using alterative refrigerants, to make comparisons and determine differences. The underlying purpose is to accelerate phase out of chemical compounds of environmental concern. The database provides bibliographic citations and abstracts for publications that may be useful in research and design of air-conditioning and refrigeration equipment. The complete documents are not included, though some may be added at a later date. The database identifies sources of specific information on various refrigerants. It addresses lubricants including alkylbenzene, polyalkylene glycol, polyolester, and other synthetics as well as mineral oils. It also references documents addressing compatibility of refrigerants and lubricants with metals, plastics, elastomers, motor insulation, and other materials used in refrigerant circuits. Incomplete citations or abstracts are provided for some documents. They are included to accelerate availability of the information and will be completed or replaced in future updates.

  12. Human health risk assessment database, "the NHSRC toxicity value database": supporting the risk assessment process at US EPA's National Homeland Security Research Center.

    Science.gov (United States)

    Moudgal, Chandrika J; Garrahan, Kevin; Brady-Roberts, Eletha; Gavrelis, Naida; Arbogast, Michelle; Dun, Sarah

    2008-11-15

    The toxicity value database of the United States Environmental Protection Agency's (EPA) National Homeland Security Research Center has been in development since 2004. The toxicity value database includes a compilation of agent property, toxicity, dose-response, and health effects data for 96 agents: 84 chemical and radiological agents and 12 biotoxins. The database is populated with multiple toxicity benchmark values and agent property information from secondary sources, with web links to the secondary sources, where available. A selected set of primary literature citations and associated dose-response data are also included. The toxicity value database offers a powerful means to quickly and efficiently gather pertinent toxicity and dose-response data for a number of agents that are of concern to the nation's security. This database, in conjunction with other tools, will play an important role in understanding human health risks, and will provide a means for risk assessors and managers to make quick and informed decisions on the potential health risks and determine appropriate responses (e.g., cleanup) to agent release. A final, stand alone MS ACESSS working version of the toxicity value database was completed in November, 2007.

  13. Human health risk assessment database, 'the NHSRC toxicity value database': Supporting the risk assessment process at US EPA's National Homeland Security Research Center

    International Nuclear Information System (INIS)

    Moudgal, Chandrika J.; Garrahan, Kevin; Brady-Roberts, Eletha; Gavrelis, Naida; Arbogast, Michelle; Dun, Sarah

    2008-01-01

    The toxicity value database of the United States Environmental Protection Agency's (EPA) National Homeland Security Research Center has been in development since 2004. The toxicity value database includes a compilation of agent property, toxicity, dose-response, and health effects data for 96 agents: 84 chemical and radiological agents and 12 biotoxins. The database is populated with multiple toxicity benchmark values and agent property information from secondary sources, with web links to the secondary sources, where available. A selected set of primary literature citations and associated dose-response data are also included. The toxicity value database offers a powerful means to quickly and efficiently gather pertinent toxicity and dose-response data for a number of agents that are of concern to the nation's security. This database, in conjunction with other tools, will play an important role in understanding human health risks, and will provide a means for risk assessors and managers to make quick and informed decisions on the potential health risks and determine appropriate responses (e.g., cleanup) to agent release. A final, stand alone MS ACESSS working version of the toxicity value database was completed in November, 2007

  14. Analysis of the environmental behavior of farmers for non-point source pollution control and management in a water source protection area in China.

    Science.gov (United States)

    Wang, Yandong; Yang, Jun; Liang, Jiping; Qiang, Yanfang; Fang, Shanqi; Gao, Minxue; Fan, Xiaoyu; Yang, Gaihe; Zhang, Baowen; Feng, Yongzhong

    2018-08-15

    The environmental behavior of farmers plays an important role in exploring the causes of non-point source pollution and taking scientific control and management measures. Based on the theory of planned behavior (TPB), the present study investigated the environmental behavior of farmers in the Water Source Area of the Middle Route of the South-to-North Water Diversion Project in China. Results showed that TPB could explain farmers' environmental behavior (SMC=0.26) and intention (SMC=0.36) well. Furthermore, the farmers' attitude towards behavior (AB), subjective norm (SN), and perceived behavioral control (PBC) positively and significantly influenced their environmental intention; their environmental intention further impacted their behavior. SN was proved to be the main key factor indirectly influencing the farmers' environmental behavior, while PBC had no significant and direct effect. Moreover, environmental knowledge following as a moderator, gender and age was used as control variables to conduct the environmental knowledge on TPB construct moderated mediation analysis. It demonstrated that gender had a significant controlling effect on environmental behavior; that is, males engage in more environmentally friendly behaviors. However, age showed a significant negative controlling effect on pro-environmental intention and an opposite effect on pro-environmental behavior. In addition, environmental knowledge could negatively moderate the relationship between PBC and environmental intention. PBC had a greater impact on the environmental intention of farmers with poor environmental knowledge, compared to those with plenty environmental knowledge. Altogether, the present study could provide a theoretical basis for non-point source pollution control and management. Copyright © 2018 Elsevier B.V. All rights reserved.

  15. Database Vs Data Warehouse

    Directory of Open Access Journals (Sweden)

    2007-01-01

    Full Text Available Data warehouse technology includes a set of concepts and methods that offer the users useful information for decision making. The necessity to build a data warehouse arises from the necessity to improve the quality of information in the organization. The date proceeding from different sources, having a variety of forms - both structured and unstructured, are filtered according to business rules and are integrated in a single large data collection. Using informatics solutions, managers have understood that data stored in operational systems - including databases, are an informational gold mine that must be exploited. Data warehouses have been developed to answer the increasing demands for complex analysis, which could not be properly achieved with operational databases. The present paper emphasizes some of the criteria that information application developers can use in order to choose between a database solution or a data warehouse one.

  16. Industrial point source CO2 emission strength estimation with aircraft measurements and dispersion modelling.

    Science.gov (United States)

    Carotenuto, Federico; Gualtieri, Giovanni; Miglietta, Franco; Riccio, Angelo; Toscano, Piero; Wohlfahrt, Georg; Gioli, Beniamino

    2018-02-22

    CO 2 remains the greenhouse gas that contributes most to anthropogenic global warming, and the evaluation of its emissions is of major interest to both research and regulatory purposes. Emission inventories generally provide quite reliable estimates of CO 2 emissions. However, because of intrinsic uncertainties associated with these estimates, it is of great importance to validate emission inventories against independent estimates. This paper describes an integrated approach combining aircraft measurements and a puff dispersion modelling framework by considering a CO 2 industrial point source, located in Biganos, France. CO 2 density measurements were obtained by applying the mass balance method, while CO 2 emission estimates were derived by implementing the CALMET/CALPUFF model chain. For the latter, three meteorological initializations were used: (i) WRF-modelled outputs initialized by ECMWF reanalyses; (ii) WRF-modelled outputs initialized by CFSR reanalyses and (iii) local in situ observations. Governmental inventorial data were used as reference for all applications. The strengths and weaknesses of the different approaches and how they affect emission estimation uncertainty were investigated. The mass balance based on aircraft measurements was quite succesful in capturing the point source emission strength (at worst with a 16% bias), while the accuracy of the dispersion modelling, markedly when using ECMWF initialization through the WRF model, was only slightly lower (estimation with an 18% bias). The analysis will help in highlighting some methodological best practices that can be used as guidelines for future experiments.

  17. Development of an updated phytoestrogen database for use with the SWAN food frequency questionnaire: intakes and food sources in a community-based, multiethnic cohort study.

    Science.gov (United States)

    Huang, Mei-Hua; Norris, Jean; Han, Weijuan; Block, Torin; Gold, Ellen; Crawford, Sybil; Greendale, Gail A

    2012-01-01

    Phytoestrogens, heterocyclic phenols found in plants, may benefit several health outcomes. However, epidemiologic studies of the health effects of dietary phytoestrogens have yielded mixed results, in part due to challenges inherent in estimating dietary intakes. The goal of this study was to improve the estimates of dietary phytoestrogen consumption using a modified Block Food Frequency Questionnaire (FFQ), a 137-item FFQ created for the Study of Women's Health Across the Nation (SWAN) in 1994. To expand the database of sources from which phytonutrient intakes were computed, we conducted a comprehensive PubMed/Medline search covering January 1994 through September 2008. The expanded database included 4 isoflavones, coumestrol, and 4 lignans. The new database estimated isoflavone content of 105 food items (76.6%) vs. 14 (10.2%) in the 1994 version and computed coumestrol content of 52 food items (38.0%), compared to 1 (0.7%) in the original version. Newly added were lignans; values for 104 FFQ food items (75.9%) were calculated. In addition, we report here the phytonutrient intakes for each racial and language group in the SWAN sample and present major food sources from which the phytonutrients came. This enhanced ascertainment of phytoestrogens will permit improved studies of their health effects.

  18. Statistical data evaluation in mobile gamma spectrometry. An optimisation of on-line search strategies in the scenario of lost point sources

    International Nuclear Information System (INIS)

    Hjerpe, T.; Samuelsson, C.

    1999-01-01

    There is a potential risk that hazardous radioactive sources could enter the environment, e.g. via satellite debris, smuggled radioactive goods or lost metal scrap. From a radiation protection point of view there is a need for rapid and reliable methods for locating and identifying sources. Car-borne and air-borne detector systems are suitable for the task. The condition in this work is a situation where the missing radionuclide is known, which is not an unlikely scenario. The possibility that the source is located near a road can be high, and thus motivating a car-borne spectrometer system. The main object is to optimise on-line statistical methods in order to achieve a high probability for locating point sources, or hot spots, and still have reasonably few false alarms from variations in the natural background radiation. Data were obtained from a car-borne 3 litres (NaI(Tl) detector and two point sources, located at various distances from the road. The nuclides used were 137 Cs and 131 I. Spectra were measured stationary on the road. From these measurements spectra we have reconstructed spectra applicable to different speed and sampling times; the time 3 seconds and 50 km/h are used in this work. The maximum distance a source can be located from the road and still be detected is estimated with four different statistical analysis methods. This distance is called the detection distance, DD. The method is applied on gross counts in the full energy peak window. For each method alarm thresholds has been calculated from background data obtained in Scania (Skaane), in the south of Sweden. The results show a 30-50% difference in DD's. With this semi-theoretical approach, the two sources could be detected from 250 m ( 137 Cs, 6GBq) and 200 m ( 131 I, 4GBq). (au)

  19. Global Mammal Parasite Database version 2.0.

    Science.gov (United States)

    Stephens, Patrick R; Pappalardo, Paula; Huang, Shan; Byers, James E; Farrell, Maxwell J; Gehman, Alyssa; Ghai, Ria R; Haas, Sarah E; Han, Barbara; Park, Andrew W; Schmidt, John P; Altizer, Sonia; Ezenwa, Vanessa O; Nunn, Charles L

    2017-05-01

    Illuminating the ecological and evolutionary dynamics of parasites is one of the most pressing issues facing modern science, and is critical for basic science, the global economy, and human health. Extremely important to this effort are data on the disease-causing organisms of wild animal hosts (including viruses, bacteria, protozoa, helminths, arthropods, and fungi). Here we present an updated version of the Global Mammal Parasite Database, a database of the parasites of wild ungulates (artiodactyls and perissodactyls), carnivores, and primates, and make it available for download as complete flat files. The updated database has more than 24,000 entries in the main data file alone, representing data from over 2700 literature sources. We include data on sampling method and sample sizes when reported, as well as both "reported" and "corrected" (i.e., standardized) binomials for each host and parasite species. Also included are current higher taxonomies and data on transmission modes used by the majority of species of parasites in the database. In the associated metadata we describe the methods used to identify sources and extract data from the primary literature, how entries were checked for errors, methods used to georeference entries, and how host and parasite taxonomies were standardized across the database. We also provide definitions of the data fields in each of the four files that users can download. © 2017 by the Ecological Society of America.

  20. A method for untriggered time-dependent searches for multiple flares from neutrino point sources

    International Nuclear Information System (INIS)

    Gora, D.; Bernardini, E.; Cruz Silva, A.H.

    2011-04-01

    A method for a time-dependent search for flaring astrophysical sources which can be potentially detected by large neutrino experiments is presented. The method uses a time-clustering algorithm combined with an unbinned likelihood procedure. By including in the likelihood function a signal term which describes the contribution of many small clusters of signal-like events, this method provides an effective way for looking for weak neutrino flares over different time-scales. The method is sensitive to an overall excess of events distributed over several flares which are not individually detectable. For standard cases (one flare) the discovery potential of the method is worse than a standard time-dependent point source analysis with unknown duration of the flare by a factor depending on the signal-to-background level. However, for flares sufficiently shorter than the total observation period, the method is more sensitive than a time-integrated analysis. (orig.)

  1. A method for untriggered time-dependent searches for multiple flares from neutrino point sources

    Energy Technology Data Exchange (ETDEWEB)

    Gora, D. [Deutsches Elektronen-Synchrotron (DESY), Zeuthen (Germany); Institute of Nuclear Physics PAN, Cracow (Poland); Bernardini, E.; Cruz Silva, A.H. [Institute of Nuclear Physics PAN, Cracow (Poland)

    2011-04-15

    A method for a time-dependent search for flaring astrophysical sources which can be potentially detected by large neutrino experiments is presented. The method uses a time-clustering algorithm combined with an unbinned likelihood procedure. By including in the likelihood function a signal term which describes the contribution of many small clusters of signal-like events, this method provides an effective way for looking for weak neutrino flares over different time-scales. The method is sensitive to an overall excess of events distributed over several flares which are not individually detectable. For standard cases (one flare) the discovery potential of the method is worse than a standard time-dependent point source analysis with unknown duration of the flare by a factor depending on the signal-to-background level. However, for flares sufficiently shorter than the total observation period, the method is more sensitive than a time-integrated analysis. (orig.)

  2. Numerical Procedure to Forecast the Tsunami Parameters from a Database of Pre-Simulated Seismic Unit Sources

    Science.gov (United States)

    Jiménez, César; Carbonel, Carlos; Rojas, Joel

    2018-04-01

    We have implemented a numerical procedure to forecast the parameters of a tsunami, such as the arrival time of the front of the first wave and the maximum wave height in real and virtual tidal stations along the Peruvian coast, with this purpose a database of pre-computed synthetic tsunami waveforms (or Green functions) was obtained from numerical simulation of seismic unit sources (dimension: 50 × 50 km2) for subduction zones from southern Chile to northern Mexico. A bathymetry resolution of 30 arc-sec (approximately 927 m) was used. The resulting tsunami waveform is obtained from the superposition of synthetic waveforms corresponding to several seismic unit sources contained within the tsunami source geometry. The numerical procedure was applied to the Chilean tsunami of April 1, 2014. The results show a very good correlation for stations with wave amplitude greater than 1 m, in the case of the Arica tide station an error (from the maximum height of the observed and simulated waveform) of 3.5% was obtained, for Callao station the error was 12% and the largest error was in Chimbote with 53.5%, however, due to the low amplitude of the Chimbote wave (<1 m), the overestimated error, in this case, is not important for evacuation purposes. The aim of the present research is tsunami early warning, where speed is required rather than accuracy, so the results should be taken as preliminary.

  3. Database on Aims and Visions in the COINCO Corridor

    DEFF Research Database (Denmark)

    2005-01-01

    This database contains aims and visions regarding overall regional development as well as more specific aims and visions related to transport and infrastructure in the Corridor Oslo-Göteborg-Copenhagen-Berlin. The sources used for this database are the most essential planning documents from Denmark...

  4. Hubble Source Catalog

    Science.gov (United States)

    Lubow, S.; Budavári, T.

    2013-10-01

    We have created an initial catalog of objects observed by the WFPC2 and ACS instruments on the Hubble Space Telescope (HST). The catalog is based on observations taken on more than 6000 visits (telescope pointings) of ACS/WFC and more than 25000 visits of WFPC2. The catalog is obtained by cross matching by position in the sky all Hubble Legacy Archive (HLA) Source Extractor source lists for these instruments. The source lists describe properties of source detections within a visit. The calculations are performed on a SQL Server database system. First we collect overlapping images into groups, e.g., Eta Car, and determine nearby (approximately matching) pairs of sources from different images within each group. We then apply a novel algorithm for improving the cross matching of pairs of sources by adjusting the astrometry of the images. Next, we combine pairwise matches into maximal sets of possible multi-source matches. We apply a greedy Bayesian method to split the maximal matches into more reliable matches. We test the accuracy of the matches by comparing the fluxes of the matched sources. The result is a set of information that ties together multiple observations of the same object. A byproduct of the catalog is greatly improved relative astrometry for many of the HST images. We also provide information on nondetections that can be used to determine dropouts. With the catalog, for the first time, one can carry out time domain, multi-wavelength studies across a large set of HST data. The catalog is publicly available. Much more can be done to expand the catalog capabilities.

  5. Report on Approaches to Database Translation. Final Report.

    Science.gov (United States)

    Gallagher, Leonard; Salazar, Sandra

    This report describes approaches to database translation (i.e., transferring data and data definitions from a source, either a database management system (DBMS) or a batch file, to a target DBMS), and recommends a method for representing the data structures of newly-proposed network and relational data models in a form suitable for database…

  6. Implementing a Dynamic Database-Driven Course Using LAMP

    Science.gov (United States)

    Laverty, Joseph Packy; Wood, David; Turchek, John

    2011-01-01

    This paper documents the formulation of a database driven open source architecture web development course. The design of a web-based curriculum faces many challenges: a) relative emphasis of client and server-side technologies, b) choice of a server-side language, and c) the cost and efficient delivery of a dynamic web development, database-driven…

  7. Database Description - Trypanosomes Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Trypanosomes Database Database Description General information of database Database name Trypanosomes Database...stitute of Genetics Research Organization of Information and Systems Yata 1111, Mishima, Shizuoka 411-8540, JAPAN E mail: Database...y Name: Trypanosoma Taxonomy ID: 5690 Taxonomy Name: Homo sapiens Taxonomy ID: 9606 Database description The... Article title: Author name(s): Journal: External Links: Original website information Database maintenance s...DB (Protein Data Bank) KEGG PATHWAY Database DrugPort Entry list Available Query search Available Web servic

  8. Volatile organic compound emissions from the oil and natural gas industry in the Uinta Basin, Utah: point sources compared to ambient air composition

    Science.gov (United States)

    Warneke, C.; Geiger, F.; Edwards, P. M.; Dube, W.; Pétron, G.; Kofler, J.; Zahn, A.; Brown, S. S.; Graus, M.; Gilman, J.; Lerner, B.; Peischl, J.; Ryerson, T. B.; de Gouw, J. A.; Roberts, J. M.

    2014-05-01

    The emissions of volatile organic compounds (VOCs) associated with oil and natural gas production in the Uinta Basin, Utah were measured at a ground site in Horse Pool and from a NOAA mobile laboratory with PTR-MS instruments. The VOC compositions in the vicinity of individual gas and oil wells and other point sources such as evaporation ponds, compressor stations and injection wells are compared to the measurements at Horse Pool. High mixing ratios of aromatics, alkanes, cycloalkanes and methanol were observed for extended periods of time and short-term spikes caused by local point sources. The mixing ratios during the time the mobile laboratory spent on the well pads were averaged. High mixing ratios were found close to all point sources, but gas wells using dry-gas collection, which means dehydration happens at the well, were clearly associated with higher mixing ratios than other wells. Another large source was the flowback pond near a recently hydraulically re-fractured gas well. The comparison of the VOC composition of the emissions from the oil and natural gas wells showed that wet gas collection wells compared well with the majority of the data at Horse Pool and that oil wells compared well with the rest of the ground site data. Oil wells on average emit heavier compounds than gas wells. The mobile laboratory measurements confirm the results from an emissions inventory: the main VOC source categories from individual point sources are dehydrators, oil and condensate tank flashing and pneumatic devices and pumps. Raw natural gas is emitted from the pneumatic devices and pumps and heavier VOC mixes from the tank flashings.

  9. Investigation on structuring the human body function database; Shintai kino database no kochiku ni kansuru chosa kenkyu

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1995-03-01

    Based on the concept of human life engineering database, a study was made to know how to technically make such a database fittable to the old people in the age-advancing society. It was then proposed that the old people`s human life engineering database should be prepared to serve for the development and design of life technology to be applied into the age-advancing society. An executive method of structuring the database was established through the `bathing` and `going out` selected as an action to be casestudied in the daily life of old people. As a result of the study, the proposal was made that the old people`s human body function database should be prepared as a R and D base for the life technology in the aged society. Based on the above proposal, a master plan was mapped out to structure this database with the concrete method studied for putting it into action. At the first investigation stage of the above study, documentation was made through utilizing the existing documentary database. Enterprises were also interviewed for the investigation. Pertaining to the function of old people, about 500 documents were extracted with many vague points not clarified yet. The investigation will restart in the next fiscal year. 4 refs., 38 figs., 30 tabs.

  10. ARTI Refrigerant Database

    Energy Technology Data Exchange (ETDEWEB)

    Calm, J.M.

    1992-11-09

    The database provides bibliographic citations and abstracts for publications that may be useful in research and design of air- conditioning and refrigeration equipment. The database identifies sources of specific information on R-32, R-123, R-124, R-125, R-134, R-134a, R-141b, R-142b, R-143a, R-152a, R-245ca, R-290 (propane), R- 717 (ammonia), ethers, and others as well as azeotropic and zeotropic and zeotropic blends of these fluids. It addresses lubricants including alkylbenzene, polyalkylene glycol, ester, and other synthetics as well as mineral oils. It also references documents on compatibility of refrigerants and lubricants with metals, plastics, elastomers, motor insulation, and other materials used in refrigerant circuits. A computerized version is available that includes retrieval software.

  11. A new model validation database for evaluating AERMOD, NRPB R91 and ADMS using krypton-85 data from BNFL Sellafield

    International Nuclear Information System (INIS)

    Hill, R.; Taylor, J.; Lowles, I.; Emmerson, K.; Parker, T.

    2004-01-01

    The emission of krypton-85 ( 85 Kr) from nuclear fuel reprocessing operations provide a classical passive tracer for the study of atmospheric dispersion. This is because of the persistence of this radioisotope in the atmosphere, due to its long radioactive halflife and inert chemistry; and the low background levels that result due to the limited number of anthropogenic sources globally. The BNFL Sellafield site in Cumbria (UK) is one of the most significant point sources of 85 Kr in the northern hemisphere, with 85 Kr being discharged from two stacks on the site, MAGNOX and THORP. Field experiments have been conducted since October 1996 using a cryogenic distillation technique (Janssens et al., 1986) to quantify the ground level concentration of 85 Kr. This paper reports on the construction of a model validation database to allow evaluation of regulatory atmospheric dispersion models using the measured 85 Kr concentrations as a tracer. The results of the database for local and regional scale dispersion are presented. (orig.)

  12. A Single Phase Doubly Grounded Semi-Z-Source Inverter for Photovoltaic (PV Systems with Maximum Power Point Tracking (MPPT

    Directory of Open Access Journals (Sweden)

    Tofael Ahmed

    2014-06-01

    Full Text Available In this paper, a single phase doubly grounded semi-Z-source inverter with maximum power point tracking (MPPT is proposed for photovoltaic (PV systems. This proposed system utilizes a single-ended primary inductor (SEPIC converter as DC-DC converter to implement the MPPT algorithm for tracking the maximum power from a PV array and a single phase semi-Z-source inverter for integrating the PV with AC power utilities. The MPPT controller utilizes a fast-converging algorithm to track the maximum power point (MPP and the semi-Z-source inverter utilizes a nonlinear SPWM to produce sinusoidal voltage at the output. The proposed system is able to track the MPP of PV arrays and produce an AC voltage at its output by utilizing only three switches. Experimental results show that the fast-converging MPPT algorithm has fast tracking response with appreciable MPP efficiency. In addition, the inverter shows the minimization of common mode leakage current with its ground sharing feature and reduction of the THD as well as DC current components at the output during DC-AC conversion.

  13. The Net Enabled Waste Management Database as an international source of radioactive waste management information

    International Nuclear Information System (INIS)

    Csullog, G.W.; Friedrich, V.; Miaw, S.T.W.; Tonkay, D.; Petoe, A.

    2002-01-01

    The IAEA's Net Enabled Waste Management Database (NEWMDB) is an integral part of the IAEA's policies and strategy related to the collection and dissemination of information, both internal to the IAEA in support of its activities and external to the IAEA (publicly available). The paper highlights the NEWMDB's role in relation to the routine reporting of status and trends in radioactive waste management, in assessing the development and implementation of national systems for radioactive waste management, in support of a newly developed indicator of sustainable development for radioactive waste management, in support of reporting requirements for the Joint Convention on the Safety of Spent Fuel Management and on the Safety of Radioactive Waste Management, in support of IAEA activities related to the harmonization of waste management information at the national and international levels and in relation to the management of spent/disused sealed radioactive sources. (author)

  14. SREM - WRS system module number 3348 for calculating the removal flux due to point, line or disc sources

    International Nuclear Information System (INIS)

    Grimstone, M.J.

    1978-06-01

    The WRS Modular Programming System has been developed as a means by which programmes may be more efficiently constructed, maintained and modified. In this system a module is a self-contained unit typically composed of one or more Fortran routines, and a programme is constructed from a number of such modules. This report describes one WRS module, the function of which is to calculate the uncollided flux and first-collision source from a disc source in a slab geometry system, a line source at the centre of a cylindrical system or a point source at the centre of a spherical system. The information given in this manual is of use both to the programmer wishing to incorporate the module in a programme, and to the user of such a programme. (author)

  15. Database Description - SKIP Stemcell Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us SKIP Stemcell Database Database Description General information of database Database name SKIP Stemcell Database...rsity Journal Search: Contact address http://www.skip.med.keio.ac.jp/en/contact/ Database classification Human Genes and Diseases Dat...abase classification Stemcell Article Organism Taxonomy Name: Homo sapiens Taxonomy ID: 9606 Database...ks: Original website information Database maintenance site Center for Medical Genetics, School of medicine, ...lable Web services Not available URL of Web services - Need for user registration Not available About This Database Database

  16. Emission & Generation Resource Integrated Database (eGRID)

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Emissions & Generation Resource Integrated Database (eGRID) is an integrated source of data on environmental characteristics of electric power generation....

  17. Database Description - Arabidopsis Phenome Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Arabidopsis Phenome Database Database Description General information of database Database n... BioResource Center Hiroshi Masuya Database classification Plant databases - Arabidopsis thaliana Organism T...axonomy Name: Arabidopsis thaliana Taxonomy ID: 3702 Database description The Arabidopsis thaliana phenome i...heir effective application. We developed the new Arabidopsis Phenome Database integrating two novel database...seful materials for their experimental research. The other, the “Database of Curated Plant Phenome” focusing

  18. Database for waste glass composition and properties

    International Nuclear Information System (INIS)

    Peters, R.D.; Chapman, C.C.; Mendel, J.E.; Williams, C.G.

    1993-09-01

    A database of waste glass composition and properties, called PNL Waste Glass Database, has been developed. The source of data is published literature and files from projects funded by the US Department of Energy. The glass data have been organized into categories and corresponding data files have been prepared. These categories are glass chemical composition, thermal properties, leaching data, waste composition, glass radionuclide composition and crystallinity data. The data files are compatible with commercial database software. Glass compositions are linked to properties across the various files using a unique glass code. Programs have been written in database software language to permit searches and retrievals of data. The database provides easy access to the vast quantities of glass compositions and properties that have been studied. It will be a tool for researchers and others investigating vitrification and glass waste forms

  19. Biosolid stockpiles are a significant point source for greenhouse gas emissions.

    Science.gov (United States)

    Majumder, Ramaprasad; Livesley, Stephen J; Gregory, David; Arndt, Stefan K

    2014-10-01

    The wastewater treatment process generates large amounts of sewage sludge that are dried and then often stored in biosolid stockpiles in treatment plants. Because the biosolids are rich in decomposable organic matter they could be a significant source for greenhouse gas (GHG) emissions, yet there are no direct measurements of GHG from stockpiles. We therefore measured the direct emissions of methane (CH4), nitrous oxide (N2O) and carbon dioxide (CO2) on a monthly basis from three different age classes of biosolid stockpiles at the Western Treatment Plant (WTP), Melbourne, Australia, from December 2009 to November 2011 using manual static chambers. All biosolid stockpiles were a significant point source for CH4 and N2O emissions. The youngest biosolids (nitrate and ammonium concentration. We also modeled CH4 emissions based on a first order decay model and the model based estimated annual CH4 emissions were higher as compared to the direct field based estimated annual CH4 emissions. Our results indicate that labile organic material in stockpiles is decomposed over time and that nitrogen decomposition processes lead to significant N2O emissions. Carbon decomposition favors CO2 over CH4 production probably because of aerobic stockpile conditions or CH4 oxidation in the outer stockpile layers. Although the GHG emission rate decreased with biosolid age, managers of biosolid stockpiles should assess alternate storage or uses for biosolids to avoid nutrient losses and GHG emissions. Copyright © 2014 Elsevier Ltd. All rights reserved.

  20. SSC lattice database and graphical interface

    International Nuclear Information System (INIS)

    Trahern, C.G.; Zhou, J.

    1991-11-01

    When completed the Superconducting Super Collider will be the world's largest accelerator complex. In order to build this system on schedule, the use of database technologies will be essential. In this paper we discuss one of the database efforts underway at the SSC, the lattice database. The SSC lattice database provides a centralized source for the design of each major component of the accelerator complex. This includes the two collider rings, the High Energy Booster, Medium Energy Booster, Low Energy Booster, and the LINAC as well as transfer and test beam lines. These designs have been created using a menagerie of programs such as SYNCH, DIMAD, MAD, TRANSPORT, MAGIC, TRACE3D AND TEAPOT. However, once a design has been completed, it is entered into a uniform database schema in the database system. In this paper we discuss the reasons for creating the lattice database and its implementation via the commercial database system SYBASE. Each lattice in the lattice database is composed of a set of tables whose data structure can describe any of the SSC accelerator lattices. In order to allow the user community access to the databases, a programmatic interface known as dbsf (for database to several formats) has been written. Dbsf creates ascii input files appropriate to the above mentioned accelerator design programs. In addition it has a binary dataset output using the Self Describing Standard data discipline provided with the Integrated Scientific Tool Kit software tools. Finally we discuss the graphical interfaces to the lattice database. The primary interface, known as OZ, is a simulation environment as well as a database browser

  1. Data integration and knowledge discovery in biomedical databases. Reliable information from unreliable sources

    Directory of Open Access Journals (Sweden)

    A Mitnitski

    2003-01-01

    Full Text Available To better understand information about human health from databases we analyzed three datasets collected for different purposes in Canada: a biomedical database of older adults, a large population survey across all adult ages, and vital statistics. Redundancy in the variables was established, and this led us to derive a generalized (macroscopic state variable, being a fitness/frailty index that reflects both individual and group health status. Evaluation of the relationship between fitness/frailty and the mortality rate revealed that the latter could be expressed in terms of variables generally available from any cross-sectional database. In practical terms, this means that the risk of mortality might readily be assessed from standard biomedical appraisals collected for other purposes.

  2. International shock-wave database project : report of the requirements workshop.

    Energy Technology Data Exchange (ETDEWEB)

    Aidun, John Bahram (Institute of Problems of chemical Physics of Russian Academy of Sciences); Lomonosov, Igor V. (Institute of Problems of chemical Physics of Russian Academy of Sciences); Levashov, Pavel R. (Joint Institute for High Temperatures of Russian Academy of Sciences)

    2012-03-01

    We report on the requirements workshop for a new project, the International Shock-Wave database (ISWdb), which was held October 31 - November 2, 2011, at GSI, Darmstadt, Germany. Participants considered the idea of this database, its structure, technical requirements, content, and principles of operation. This report presents the consensus conclusions from the workshop, key discussion points, and the goals and plan for near-term and intermediate-term development of the ISWdb. The main points of consensus from the workshop were: (1) This international database is of interest and of practical use for the shock-wave and high pressure physics communities; (2) Intermediate state information and off-Hugoniot information is important and should be included in ISWdb; (3) Other relevant high pressure and auxiliary data should be included to the database, in the future; (4) Information on the ISWdb needs to be communicated, broadly, to the research community; and (5) Operating structure will consist of an Advisory Board, subject-matter expert Moderators to vet submitted data, and the database Project Team. This brief report is intended to inform the shock-wave research community and interested funding agencies about the project, as its success, ultimately, depends on both of these groups finding sufficient value in the database to use it, contribute to it, and support it.

  3. Distributed Pseudo-Random Number Generation and Its Application to Cloud Database

    OpenAIRE

    Chen, Jiageng; Miyaji, Atsuko; Su, Chunhua

    2014-01-01

    Cloud database is now a rapidly growing trend in cloud computing market recently. It enables the clients run their computation on out-sourcing databases or access to some distributed database service on the cloud. At the same time, the security and privacy concerns is major challenge for cloud database to continue growing. To enhance the security and privacy of the cloud database technology, the pseudo-random number generation (PRNG) plays an important roles in data encryptions and privacy-pr...

  4. Overview of Historical Earthquake Document Database in Japan and Future Development

    Science.gov (United States)

    Nishiyama, A.; Satake, K.

    2014-12-01

    In Japan, damage and disasters from historical large earthquakes have been documented and preserved. Compilation of historical earthquake documents started in the early 20th century and 33 volumes of historical document source books (about 27,000 pages) have been published. However, these source books are not effectively utilized for researchers due to a contamination of low-reliability historical records and a difficulty for keyword searching by characters and dates. To overcome these problems and to promote historical earthquake studies in Japan, construction of text database started in the 21 century. As for historical earthquakes from the beginning of the 7th century to the early 17th century, "Online Database of Historical Documents in Japanese Earthquakes and Eruptions in the Ancient and Medieval Ages" (Ishibashi, 2009) has been already constructed. They investigated the source books or original texts of historical literature, emended the descriptions, and assigned the reliability of each historical document on the basis of written age. Another database compiled the historical documents for seven damaging earthquakes occurred along the Sea of Japan coast in Honshu, central Japan in the Edo period (from the beginning of the 17th century to the middle of the 19th century) and constructed text database and seismic intensity data base. These are now publicized on the web (written only in Japanese). However, only about 9 % of the earthquake source books have been digitized so far. Therefore, we plan to digitize all of the remaining historical documents by the research-program which started in 2014. The specification of the data base will be similar for previous ones. We also plan to combine this database with liquefaction traces database, which will be constructed by other research program, by adding the location information described in historical documents. Constructed database would be utilized to estimate the distributions of seismic intensities and tsunami

  5. PostgreSQL in the database landscape

    CERN Multimedia

    CERN. Geneva; Riggs, Simon

    2013-01-01

    This presentation targets the exposure of PostgreSQL and its main highlights in two parts: PostgreSQL today, by Harald Armin Massa This will explore the functionalities and capabilities of PostgreSQL; point out differences to other available databases; give information about the PostgreSQL project how it ensures the quality of this software. PostgreSQL and Extremely Large Databases, by Simon Riggs presenting an outlook on what is happening with PostgreSQL and Extremely Large Databases. About the speakers Simon Riggs is founder and CTO of 2ndQuadrant. He is working in the AXLE project. He works as an Architect and Developer of new features for PostgreSQL, setting technical directions for 2ndQuadrant and as a Database Systems Architect for 2ndQuadrant customers. Simon is the author of PostgreSQL 9 Admin Cookbook; and a committer to the PostgreSQL project. Harald Armin Massa studied computers and economics, he's self employed since 1999, doing software development in Python and ...

  6. Quality Assurance Source Requirements Traceability Database

    International Nuclear Information System (INIS)

    MURTHY, R.; NAYDENOVA, A.; DEKLEVER, R.; BOONE, A.

    2006-01-01

    At the Yucca Mountain Project the Project Requirements Processing System assists in the management of relationships between regulatory and national/industry standards source criteria, and Quality Assurance Requirements and Description document (DOE/R W-0333P) requirements to create compliance matrices representing respective relationships. The matrices are submitted to the U.S. Nuclear Regulatory Commission to assist in the commission's review, interpretation, and concurrence with the Yucca Mountain Project QA program document. The tool is highly customized to meet the needs of the Office of Civilian Radioactive Waste Management Office of Quality Assurance

  7. Recent Advances and Coming Attractions in the NASA/IPAC Extragalactic Database

    Science.gov (United States)

    Mazzarella, Joseph M.; Baker, Kay; Pan Chan, Hiu; Chen, Xi; Ebert, Rick; Frayer, Cren; Helou, George; Jacobson, Jeffery D.; Lo, Tak M.; Madore, Barry; Ogle, Patrick M.; Pevunova, Olga; Steer, Ian; Schmitz, Marion; Terek, Scott

    2017-01-01

    We review highlights of recent advances and developments underway at the NASA/IPAC Extragalactic Database (NED). Extensive updates have been made to the infrastructure and processes essential for scaling NED for the next steps in its evolution. A major overhaul of the data integration pipeline provides greater modularity and parallelization to increase the rate of source cross-matching and data integration. The new pipeline was used recently to fold in data for nearly 300,000 sources published in over 900 recent journal articles, as well as fundamental parameters for 42 million sources in the Spitzer Enhanced Imaging Products Source List. The latter has added over 360 million photometric measurements at 3.6, 4.5, 5.8. 8.0 (IRAC) and 24 microns (MIPS) to the spectral energy distributions of affected objects in NED. The recent discovery of super-luminous spiral galaxies (Ogle et al. 2016) exemplifies the opportunities for science discovery and data mining available directly from NED’s unique data synthesis, spanning the spectrum from gamma ray through radio frequencies. The number of references in NED has surpassed 103,000. In the coming year, cross-identifications of sources in the 2MASS Point Source Catalog and in the AllWISE Source Catalog with prior objects in the database (including GALEX) will increase the holdings to over a billion distinct objects, providing a rich resource for multi-wavelength analysis. Information about a recent surge in growth of redshift-independent distances in NED is presented at this meeting by Steer et al. (2017). Website updates include a ’simple search’ to perform common queries in a single entry field, an interface to query the image repository with options to sort and filter the initial results, connectivity to the IRSA Finder Chart service, as well as a program interface to query images using the international virtual observatory Simple Image Access protocol. Graphical characterizations of NED content and completeness are

  8. Carbon Dioxide Capture and Separation Techniques for Gasification-based Power Generation Point Sources

    Energy Technology Data Exchange (ETDEWEB)

    Pennline, H.W.; Luebke, D.R.; Jones, K.L.; Morsi, B.I. (Univ. of Pittsburgh, PA); Heintz, Y.J. (Univ. of Pittsburgh, PA); Ilconich, J.B. (Parsons)

    2007-06-01

    The capture/separation step for carbon dioxide (CO2) from large-point sources is a critical one with respect to the technical feasibility and cost of the overall carbon sequestration scenario. For large-point sources, such as those found in power generation, the carbon dioxide capture techniques being investigated by the in-house research area of the National Energy Technology Laboratory possess the potential for improved efficiency and reduced costs as compared to more conventional technologies. The investigated techniques can have wide applications, but the research has focused on capture/separation of carbon dioxide from flue gas (post-combustion from fossil fuel-fired combustors) and from fuel gas (precombustion, such as integrated gasification combined cycle or IGCC). With respect to fuel gas applications, novel concepts are being developed in wet scrubbing with physical absorption; chemical absorption with solid sorbents; and separation by membranes. In one concept, a wet scrubbing technique is being investigated that uses a physical solvent process to remove CO2 from fuel gas of an IGCC system at elevated temperature and pressure. The need to define an ideal solvent has led to the study of the solubility and mass transfer properties of various solvents. Pertaining to another separation technology, fabrication techniques and mechanistic studies for membranes separating CO2 from the fuel gas produced by coal gasification are also being performed. Membranes that consist of CO2-philic ionic liquids encapsulated into a polymeric substrate have been investigated for permeability and selectivity. Finally, dry, regenerable processes based on sorbents are additional techniques for CO2 capture from fuel gas. An overview of these novel techniques is presented along with a research progress status of technologies related to membranes and physical solvents.

  9. Search for cosmic γ-point sources above 20 TeV with the HEGRA detector fields

    International Nuclear Information System (INIS)

    Prahl, J.

    1999-01-01

    The topic of this thesis is the search for point sources of high energy γ-rays (above 20 TeV) during the time interval from December 1993 until September 1995 with the HEGRA scintillator and AIROBICC arrays. The main aim is to exploit the available data as complete as possible in order to achieve a maximum in sensitivity. A thorough MtCarlo simulation of the registered events from the cosmic-ray-induced air showers allows for the use of all acquired events [de

  10. Two- and three-particle interference correlations of identical bosons and fermions with close momenta in the model of independent point-like sources

    International Nuclear Information System (INIS)

    Lyuboshits, V.L.

    1991-01-01

    Interference correlations introduced between identical particles with close momenta by the effect of Bose or Fermi statistics are discussed. Relations describing two- and three-particle correlations of identical bosons and fermions with arbitrary spin and arbitrary spin polarization are obtained on the basis of the model of independent single-particle point-like sources. The general structure of the dependence of narrow two- and three-particle correlations on the difference of the four-momenta in the presence of several groups of single-particle sources with different space-time distributions is analyzed. The idea of many-particle point sources of identical bosons is introduced. The suppression of two- and three-particle interference correlations between identical π mesons under conditions when one or several many-particle sources are added to a system of randomly distributed independent single-particle sources is studied. It is shown that if the multiplicities of the particles emitted by the sources are distributed according to the Poisson law, the present results agree with the relations obtained by means of the formalism of coherent states. This agreement also holds in the limit of very large multiplicities with any distribution laws

  11. Designing a database for performance assessment: Lessons learned from WIPP

    International Nuclear Information System (INIS)

    Martell, M.A.; Schenker, A.

    1997-01-01

    The Waste Isolation Pilot Plant (WIPP) Compliance Certification Application (CCA) Performance Assessment (PA) used a relational database that was originally designed only to supply the input parameters required for implementation of the PA codes. Reviewers used the database as a point of entry to audit quality assurance measures for control, traceability, and retrievability of input information used for analysis, and output/work products. During these audits it became apparent that modifications to the architecture and scope of the database would benefit the EPA regulator and other stakeholders when reviewing the recertification application. This paper contains a discussion of the WPP PA CCA database and lessons learned for designing a database

  12. An Open-source Toolbox for Analysing and Processing PhysioNet Databases in MATLAB and Octave.

    Science.gov (United States)

    Silva, Ikaro; Moody, George B

    The WaveForm DataBase (WFDB) Toolbox for MATLAB/Octave enables integrated access to PhysioNet's software and databases. Using the WFDB Toolbox for MATLAB/Octave, users have access to over 50 physiological databases in PhysioNet. The toolbox provides access over 4 TB of biomedical signals including ECG, EEG, EMG, and PLETH. Additionally, most signals are accompanied by metadata such as medical annotations of clinical events: arrhythmias, sleep stages, seizures, hypotensive episodes, etc. Users of this toolbox should easily be able to reproduce, validate, and compare results published based on PhysioNet's software and databases.

  13. SAADA: Astronomical Databases Made Easier

    Science.gov (United States)

    Michel, L.; Nguyen, H. N.; Motch, C.

    2005-12-01

    Many astronomers wish to share datasets with their community but have not enough manpower to develop databases having the functionalities required for high-level scientific applications. The SAADA project aims at automatizing the creation and deployment process of such databases. A generic but scientifically relevant data model has been designed which allows one to build databases by providing only a limited number of product mapping rules. Databases created by SAADA rely on a relational database supporting JDBC and covered by a Java layer including a lot of generated code. Such databases can simultaneously host spectra, images, source lists and plots. Data are grouped in user defined collections whose content can be seen as one unique set per data type even if their formats differ. Datasets can be correlated one with each other using qualified links. These links help, for example, to handle the nature of a cross-identification (e.g., a distance or a likelihood) or to describe their scientific content (e.g., by associating a spectrum to a catalog entry). The SAADA query engine is based on a language well suited to the data model which can handle constraints on linked data, in addition to classical astronomical queries. These constraints can be applied on the linked objects (number, class and attributes) and/or on the link qualifier values. Databases created by SAADA are accessed through a rich WEB interface or a Java API. We are currently developing an inter-operability module implanting VO protocols.

  14. The Freight Analysis Framework Verson 4 (FAF4) - Building the FAF4 Regional Database: Data Sources and Estimation Methodologies

    Energy Technology Data Exchange (ETDEWEB)

    Hwang, Ho-Ling [ORNL; Hargrove, Stephanie [ORNL; Chin, Shih-Miao [ORNL; Wilson, Daniel W [ORNL; Taylor, Rob D [ORNL; Davidson, Diane [ORNL

    2016-09-01

    The Freight Analysis Framework (FAF) integrates data from a variety of sources to create a comprehensive national picture of freight movements among states and major metropolitan areas by all modes of transportation. It provides a national picture of current freight flows to, from, and within the United States, assigns the flows to the transportation network, and projects freight flow patterns into the future. The FAF4 is the fourth database of its kind, FAF1 provided estimates for truck, rail, and water tonnage for calendar year 1998, FAF2 provided a more complete picture based on the 2002 Commodity Flow Survey (CFS) and FAF3 made further improvements building on the 2007 CFS. Since the first FAF effort, a number of changes in both data sources and products have taken place. The FAF4 flow matrix described in this report is used as the base-year data to forecast future freight activities, projecting shipment weights and values from year 2020 to 2045 in five-year intervals. It also provides the basis for annual estimates to the FAF4 flow matrix, aiming to provide users with the timeliest data. Furthermore, FAF4 truck freight is routed on the national highway network to produce the FAF4 network database and flow assignments for truck. This report details the data sources and methodologies applied to develop the base year 2012 FAF4 database. An overview of the FAF4 components is briefly discussed in Section 2. Effects on FAF4 from the changes in the 2012 CFS are highlighted in Section 3. Section 4 provides a general discussion on the process used in filling data gaps within the domestic CFS matrix, specifically on the estimation of CFS suppressed/unpublished cells. Over a dozen CFS OOS components of FAF4 are then addressed in Section 5 through Section 11 of this report. This includes discussions of farm-based agricultural shipments in Section 5, shipments from fishery and logging sectors in Section 6. Shipments of municipal solid wastes and debris from construction

  15. CardioTF, a database of deconstructing transcriptional circuits in the heart system.

    Science.gov (United States)

    Zhen, Yisong

    2016-01-01

    Information on cardiovascular gene transcription is fragmented and far behind the present requirements of the systems biology field. To create a comprehensive source of data for cardiovascular gene regulation and to facilitate a deeper understanding of genomic data, the CardioTF database was constructed. The purpose of this database is to collate information on cardiovascular transcription factors (TFs), position weight matrices (PWMs), and enhancer sequences discovered using the ChIP-seq method. The Naïve-Bayes algorithm was used to classify literature and identify all PubMed abstracts on cardiovascular development. The natural language learning tool GNAT was then used to identify corresponding gene names embedded within these abstracts. Local Perl scripts were used to integrate and dump data from public databases into the MariaDB management system (MySQL). In-house R scripts were written to analyze and visualize the results. Known cardiovascular TFs from humans and human homologs from fly, Ciona, zebrafish, frog, chicken, and mouse were identified and deposited in the database. PWMs from Jaspar, hPDI, and UniPROBE databases were deposited in the database and can be retrieved using their corresponding TF names. Gene enhancer regions from various sources of ChIP-seq data were deposited into the database and were able to be visualized by graphical output. Besides biocuration, mouse homologs of the 81 core cardiac TFs were selected using a Naïve-Bayes approach and then by intersecting four independent data sources: RNA profiling, expert annotation, PubMed abstracts and phenotype. The CardioTF database can be used as a portal to construct transcriptional network of cardiac development. Database URL: http://www.cardiosignal.org/database/cardiotf.html.

  16. A high-energy nuclear database proposal

    International Nuclear Information System (INIS)

    Brown, D.A.; Vogt, R.; UC Davis, CA

    2006-01-01

    We propose to develop a high-energy heavy-ion experimental database and make it accessible to the scientific community through an on-line interface. This database will be searchable and cross-indexed with relevant publications, including published detector descriptions. Since this database will be a community resource, it requires the high-energy nuclear physics community's financial and manpower support. This database should eventually contain all published data from the Bevalac, AGS and SPS to RHIC and LHC energies, proton-proton to nucleus-nucleus collisions as well as other relevant systems, and all measured observables. Such a database would have tremendous scientific payoff as it makes systematic studies easier and allows simpler benchmarking of theoretical models to a broad range of old and new experiments. Furthermore, there is a growing need for compilations of high-energy nuclear data for applications including stockpile stewardship, technology development for inertial confinement fusion and target and source development for upcoming facilities such as the Next Linear Collider. To enhance the utility of this database, we propose periodically performing evaluations of the data and summarizing the results in topical reviews. (author)

  17. Use of electronic healthcare records in large-scale simple randomized trials at the point of care for the documentation of value-based medicine.

    Science.gov (United States)

    van Staa, T-P; Klungel, O; Smeeth, L

    2014-06-01

    A solid foundation of evidence of the effects of an intervention is a prerequisite of evidence-based medicine. The best source of such evidence is considered to be randomized trials, which are able to avoid confounding. However, they may not always estimate effectiveness in clinical practice. Databases that collate anonymized electronic health records (EHRs) from different clinical centres have been widely used for many years in observational studies. Randomized point-of-care trials have been initiated recently to recruit and follow patients using the data from EHR databases. In this review, we describe how EHR databases can be used for conducting large-scale simple trials and discuss the advantages and disadvantages of their use. © 2014 The Association for the Publication of the Journal of Internal Medicine.

  18. Hardware-accelerated Point Generation and Rendering of Point-based Impostors

    DEFF Research Database (Denmark)

    Bærentzen, Jakob Andreas

    2005-01-01

    This paper presents a novel scheme for generating points from triangle models. The method is fast and lends itself well to implementation using graphics hardware. The triangle to point conversion is done by rendering the models, and the rendering may be performed procedurally or by a black box API....... I describe the technique in detail and discuss how the generated point sets can easily be used as impostors for the original triangle models used to create the points. Since the points reside solely in GPU memory, these impostors are fairly efficient. Source code is available online....

  19. An Open-source Toolbox for Analysing and Processing PhysioNet Databases in MATLAB and Octave

    Directory of Open Access Journals (Sweden)

    Ikaro Silva

    2014-09-01

    Full Text Available The WaveForm DataBase (WFDB Toolbox for MATLAB/Octave enables  integrated access to PhysioNet's software and databases. Using the WFDB Toolbox for MATLAB/Octave, users have access to over 50 physiological databases in PhysioNet. The toolbox allows direct loading into MATLAB/Octave's workspace of over 4 TB of biomedical signals including ECG, EEG, EMG, and PLETH. Additionally, most signals are accompanied by meta data such as medical annotations of clinical events: arrhythmias, sleep stages, seizures, hypotensive episodes, etc. Users of this toolbox should easily be able to reproduce, validate, and compare results published based on PhysioNet's software and databases.

  20. First combined search for neutrino point-sources in the southern sky with the ANTARES and IceCube neutrino telescopes

    Directory of Open Access Journals (Sweden)

    Barrios-Martí J.

    2016-01-01

    Full Text Available A search for cosmic neutrino point-like sources using the ANTARES and IceCube neutrino telescopes over the Southern Hemisphere is presented. The ANTARES data were collected between January 2007 and December 2012, whereas the IceCube data ranges from April 2008 to May 2011. An unbinned maximum likelihood method is used to search for a localized excess of muon events in the southern sky assuming an E−2 neutrino source spectrum. A search over a pre-selected list of candidate sources has also been carried out for different source assumptions: spectral indices of 2.0 and 2.5, and energy cutoffs of 1 PeV, 300 TeV and 100 TeV. No significant excess over the background has been found, and upper limits for the candidate sources are presented compared to the individual experiments.

  1. AN IN-DEPTH VIEW OF THE MID-INFRARED PROPERTIES OF POINT SOURCES AND THE DIFFUSE ISM IN THE SMC GIANT H II REGION, N66

    International Nuclear Information System (INIS)

    Whelan, David G.; Johnson, Kelsey E.; Indebetouw, Rémy; Lebouteiller, Vianney; Galliano, Frédéric; Peeters, Els; Bernard-Salas, Jeronimo; Brandl, Bernhard R.

    2013-01-01

    The focus of this work is to study mid-infrared point sources and the diffuse interstellar medium (ISM) in the low-metallicity (∼0.2 Z ☉ ) giant H II region N66 in order to determine properties that may shed light on star formation in these conditions. Using the Spitzer Space Telescope's Infrared Spectrograph, we study polycyclic aromatic hydrocarbon (PAH), dust continuum, silicate, and ionic line emission from 14 targeted infrared point sources as well as spectra of the diffuse ISM that is representative of both the photodissociation regions (PDRs) and the H II regions. Among the point source spectra, we spectroscopically confirm that the brightest mid-infrared point source is a massive embedded young stellar object, we detect silicates in emission associated with two young stellar clusters, and we see spectral features of a known B[e] star that are commonly associated with Herbig Be stars. In the diffuse ISM, we provide additional evidence that the very small grain population is being photodestroyed in the hard radiation field. The 11.3 μm PAH complex emission exhibits an unexplained centroid shift in both the point source and ISM spectra that should be investigated at higher signal-to-noise and resolution. Unlike studies of other regions, the 6.2 μm and 7.7 μm band fluxes are decoupled; the data points cover a large range of I 7.7 /I 11.3 PAH ratio values within a narrow band of I 6.2 /I 11.3 ratio values. Furthermore, there is a spread in PAH ionization, being more neutral in the dense PDR where the radiation field is relatively soft, but ionized in the diffuse ISM/PDR. By contrast, the PAH size distribution appears to be independent of local ionization state. Important to unresolved studies of extragalactic low-metallicity star-forming regions, we find that emission from the infrared-bright point sources accounts for only 20%-35% of the PAH emission from the entire region. These results make a comparative data set to other star-forming regions with

  2. Open Geoscience Database

    Science.gov (United States)

    Bashev, A.

    2012-04-01

    Currently there is an enormous amount of various geoscience databases. Unfortunately the only users of the majority of the databases are their elaborators. There are several reasons for that: incompaitability, specificity of tasks and objects and so on. However the main obstacles for wide usage of geoscience databases are complexity for elaborators and complication for users. The complexity of architecture leads to high costs that block the public access. The complication prevents users from understanding when and how to use the database. Only databases, associated with GoogleMaps don't have these drawbacks, but they could be hardly named "geoscience" Nevertheless, open and simple geoscience database is necessary at least for educational purposes (see our abstract for ESSI20/EOS12). We developed a database and web interface to work with them and now it is accessible at maps.sch192.ru. In this database a result is a value of a parameter (no matter which) in a station with a certain position, associated with metadata: the date when the result was obtained; the type of a station (lake, soil etc); the contributor that sent the result. Each contributor has its own profile, that allows to estimate the reliability of the data. The results can be represented on GoogleMaps space image as a point in a certain position, coloured according to the value of the parameter. There are default colour scales and each registered user can create the own scale. The results can be also extracted in *.csv file. For both types of representation one could select the data by date, object type, parameter type, area and contributor. The data are uploaded in *.csv format: Name of the station; Lattitude(dd.dddddd); Longitude(ddd.dddddd); Station type; Parameter type; Parameter value; Date(yyyy-mm-dd). The contributor is recognised while entering. This is the minimal set of features that is required to connect a value of a parameter with a position and see the results. All the complicated data

  3. The Central Point Source in G76.9++1.0 V. R. Marthi1,∗ , J. N. ...

    Indian Academy of Sciences (India)

    Astr. (2011) 32, 451–455 c Indian Academy of Sciences. The Central Point Source in G76.9++1.0. V. R. Marthi1,∗. , J. N. Chengalur1, Y. Gupta1 ... emission has indeed been seen at 2 GHz with the Green Bank Telescope. (GBT), establishing the fact that scattering is responsible for its non- detection at low radio frequencies.

  4. The simulation research of dissolved nitrogen and phosphorus non-point source pollution in Xiao-Jiang watershed of Three Gorges Reservoir area.

    Science.gov (United States)

    Wu, Lei; Long, Tian-Yu; Li, Chong-Ming

    2010-01-01

    Xiao-jiang, with a basin area of almost 5,276 km(2) and a length of 182.4 km, is located in the center of the Three Gorges Reservoir Area, and is the largest tributary of the central section in Three Gorges Reservoir Area, farmland accounts for a large proportion of Xiao-jiang watershed, and the hilly cropland of purple soil is much of the farmland of the watershed. After the second phase of water storage in the Three Gorges Reservoir, the majority of sub-rivers in the reservoir area experienced eutrophication phenomenon frequently, and non-point source (NPS) pollution has become an important source of pollution in Xiao-jiang Watershed. Because dissolved nitrogen and phosphorus non-point source pollution are related to surface runoff and interflow, using climatic, topographic and land cover data from the internet and research institutes, the Semi-Distributed Land-use Runoff Process (SLURP) hydrological model was introduced to simulate the complete hydrological cycle of the Xiao-jiang Watershed. Based on the SLURP distributed hydrological model, non-point source pollution annual output load models of land use and rural residents were respectively established. Therefore, using GIS technology, considering the losses of dissolved nitrogen and phosphorus in the course of transport, a dissolved non-point source pollution load dynamic model was established by the organic coupling of the SLURP hydrological model and land-use output model. Through the above dynamic model, the annual dissolved non-point source nitrogen and phosphorus pollution output as well as the load in different types were simulated and quantitatively estimated from 2001 to 2008, furthermore, the loads of Xiao-jiang Watershed were calculated and expressed by temporal and spatial distribution in the Three Gorges Reservoir Area. The simulation results show that: the temporal changes of dissolved nitrogen and phosphorus load in the watershed are close to the inter-annual changes of rainfall runoff, and the

  5. Characteristics of infrared point sources associated with OH masers

    International Nuclear Information System (INIS)

    Mu Jimang; Esimbek, Jarken; Zhou Jianjun; Zhang Haijuan

    2010-01-01

    We collect 3249 OH maser sources from the literature published up to April 2007, and compile a new catalog of OH masers. We look for the exciting sources of these masers and their infrared properties from IRAS and MSX data, and make a statistical study. MSX sources associated with stellar 1612 MHz OH masers are located mainly above the blackbody line; this is caused by the dust absorption of stellar envelopes, especially in the MSX A band. The mid-IR sources associated with stellar OH masers are concentrated in a small region in an [A]-[D] vs. [A]-[E] diagram with a small fraction of contamination; this gives us a new criterion to search for new stellar OH masers and distinguish stellar masers from unknown types of OH masers. IR sources associated with 1612 MHz stellar OH masers show an expected result: the average flux of sources with F60 > F25 increases with increasing wavelength, while those with F60 F25.

  6. Digital Dental X-ray Database for Caries Screening

    Science.gov (United States)

    Rad, Abdolvahab Ehsani; Rahim, Mohd Shafry Mohd; Rehman, Amjad; Saba, Tanzila

    2016-06-01

    Standard database is the essential requirement to compare the performance of image analysis techniques. Hence the main issue in dental image analysis is the lack of available image database which is provided in this paper. Periapical dental X-ray images which are suitable for any analysis and approved by many dental experts are collected. This type of dental radiograph imaging is common and inexpensive, which is normally used for dental disease diagnosis and abnormalities detection. Database contains 120 various Periapical X-ray images from top to bottom jaw. Dental digital database is constructed to provide the source for researchers to use and compare the image analysis techniques and improve or manipulate the performance of each technique.

  7. One-way data transfer for PLC to VME status reporting at the Advanced Photon Source

    International Nuclear Information System (INIS)

    Stein, S.J.

    1993-01-01

    The Personnel Safety System for the experimental beamlines at the Advanced Photon Source will use a large number of Allen Bradley Programmable Logic Controllers (PLC) to replace conventional relay logic. PLCs allow for the design of a very advanced safety system that can handle a large number of I/O points. Certain situations Require monitoring of the safety system from various locations around the storage ring via EPICS OPI (operator interface)consoles. This presentation covers the method of choice for transferring data from the Personnel Safety System into an EPICS database. Specifics on PLC ladder design, EPICS database design, and hardware selection are also discussed

  8. Identification of environmentally relevant chemicals in bibliographic databases: a comparative analysis

    DEFF Research Database (Denmark)

    Ellegaard, Ole; Wallin, Johan Albert

    2013-01-01

    takes as its starting point environmentally important chemicals and the retrieval of selectively chosen substances in the four databases: SciFinder, Web of Science (WoS), Scopus and Google Scholar. The way chemical data are stored in the databases plays a major role in the recovery process...

  9. Proposal for a High Energy Nuclear Database

    International Nuclear Information System (INIS)

    Brown, David A.; Vogt, Ramona

    2005-01-01

    We propose to develop a high-energy heavy-ion experimental database and make it accessible to the scientific community through an on-line interface. This database will be searchable and cross-indexed with relevant publications, including published detector descriptions. Since this database will be a community resource, it requires the high-energy nuclear physics community's financial and manpower support. This database should eventually contain all published data from Bevalac and AGS to RHIC to CERN-LHC energies, proton-proton to nucleus-nucleus collisions as well as other relevant systems, and all measured observables. Such a database would have tremendous scientific payoff as it makes systematic studies easier and allows simpler benchmarking of theoretical models to a broad range of old and new experiments. Furthermore, there is a growing need for compilations of high-energy nuclear data for applications including stockpile stewardship, technology development for inertial confinement fusion and target and source development for upcoming facilities such as the Next Linear Collider. To enhance the utility of this database, we propose periodically performing evaluations of the data and summarizing the results in topical reviews

  10. [Access to health information sources in Spain. how to combat "infoxication"].

    Science.gov (United States)

    Navas-Martin, Miguel Ángel; Albornos-Muñoz, Laura; Escandell-García, Cintia

    2012-01-01

    Internet has become a priceless source for finding health information for both patients and healthcare professionals. However, the universality and the abundance of information can lead to unfounded conclusions about health issues that can confuse further than clarify the health information. This aspect causes intoxication of information: infoxication. The question lies in knowing how to filter the information that is useful, accurate and relevant for our purposes. In this regard, integrative portals, such as the Biblioteca Virtual de Salud, compile information at different levels (international, national and regional), different types of resources (databases, repositories, bibliographic sources, etc.), becoming a starting point for obtaining quality information. Copyright © 2011 Elsevier España, S.L. All rights reserved.

  11. Euromet action 428: transfer of ge detectors efficiency calibration from point source geometry to other geometries; Action euromet 428: transfert de l'etalonnage en rendement de detecteurs au germanium pour une source ponctuelle vers d'autres geometries

    Energy Technology Data Exchange (ETDEWEB)

    Lepy, M.Ch

    2000-07-01

    The EUROMET project 428 examines efficiency transfer computation for Ge gamma-ray spectrometers when the efficiency is known for a reference point source geometry in the 60 keV to 2 MeV energy range. For this, different methods are used, such as Monte Carlo simulation or semi-empirical computation. The exercise compares the application of these methods to the same selected experimental cases to determine the usage limitations versus the requested accuracy. For carefully examining these results and trying to derive information for improving the computation codes, this study was limited to a few simple cases, from an experimental efficiency calibration for point source at 10-cm source-to-detector distance. The first part concerns the simplest case of geometry transfer, i.e., using point sources for 3 source-to-detector distances: 2,5 and 20 cm; the second part deals with transfer from point source geometry to cylindrical geometry with three different matrices. The general results show that the deviations between the computed results and the measured efficiencies are for the most part within 10%. The quality of the results is rather inhomogeneous and shows that these codes cannot be used directly for metrological purposes. However, most of them are operational for routine measurements when efficiency uncertainties of 5-10% can be sufficient. (author)

  12. A multidisciplinary database for geophysical time series management

    Science.gov (United States)

    Montalto, P.; Aliotta, M.; Cassisi, C.; Prestifilippo, M.; Cannata, A.

    2013-12-01

    The variables collected by a sensor network constitute a heterogeneous data source that needs to be properly organized in order to be used in research and geophysical monitoring. With the time series term we refer to a set of observations of a given phenomenon acquired sequentially in time. When the time intervals are equally spaced one speaks of period or sampling frequency. Our work describes in detail a possible methodology for storage and management of time series using a specific data structure. We designed a framework, hereinafter called TSDSystem (Time Series Database System), in order to acquire time series from different data sources and standardize them within a relational database. The operation of standardization provides the ability to perform operations, such as query and visualization, of many measures synchronizing them using a common time scale. The proposed architecture follows a multiple layer paradigm (Loaders layer, Database layer and Business Logic layer). Each layer is specialized in performing particular operations for the reorganization and archiving of data from different sources such as ASCII, Excel, ODBC (Open DataBase Connectivity), file accessible from the Internet (web pages, XML). In particular, the loader layer performs a security check of the working status of each running software through an heartbeat system, in order to automate the discovery of acquisition issues and other warning conditions. Although our system has to manage huge amounts of data, performance is guaranteed by using a smart partitioning table strategy, that keeps balanced the percentage of data stored in each database table. TSDSystem also contains modules for the visualization of acquired data, that provide the possibility to query different time series on a specified time range, or follow the realtime signal acquisition, according to a data access policy from the users.

  13. Estates, bequests, and inheritances in Sweden - A look into the Belinda databases

    OpenAIRE

    Elinder, Mikael; Erixson, Oscar; Escobar, Sebastian; Ohlsson, Henry

    2014-01-01

    The objective of this paper is to describe two new administrative Swedish databases, referred to as the Belinda databases. Together, these databases contain the most detailed individual-level data on estates, bequests, and inheritances currently available. We present descriptive statistics for the key variables in the databases to give a picture of the size of estates, the content of the bequests, and who the recipients of the inheritances are. The statistics may serve as a point of reference...

  14. Using Large Diabetes Databases for Research.

    Science.gov (United States)

    Wild, Sarah; Fischbacher, Colin; McKnight, John

    2016-09-01

    There are an increasing number of clinical, administrative and trial databases that can be used for research. These are particularly valuable if there are opportunities for linkage to other databases. This paper describes examples of the use of large diabetes databases for research. It reviews the advantages and disadvantages of using large diabetes databases for research and suggests solutions for some challenges. Large, high-quality databases offer potential sources of information for research at relatively low cost. Fundamental issues for using databases for research are the completeness of capture of cases within the population and time period of interest and accuracy of the diagnosis of diabetes and outcomes of interest. The extent to which people included in the database are representative should be considered if the database is not population based and there is the intention to extrapolate findings to the wider diabetes population. Information on key variables such as date of diagnosis or duration of diabetes may not be available at all, may be inaccurate or may contain a large amount of missing data. Information on key confounding factors is rarely available for the nondiabetic or general population limiting comparisons with the population of people with diabetes. However comparisons that allow for differences in distribution of important demographic factors may be feasible using data for the whole population or a matched cohort study design. In summary, diabetes databases can be used to address important research questions. Understanding the strengths and limitations of this approach is crucial to interpret the findings appropriately. © 2016 Diabetes Technology Society.

  15. Integration of population census and water point mapping data-A case study of Cambodia, Liberia and Tanzania.

    Science.gov (United States)

    Yu, Weiyu; Wardrop, Nicola A; Bain, Robert; Wright, Jim A

    2017-07-01

    Sustainable Development Goal (SDG) 6 has expanded the Millennium Development Goals' focus from improved drinking-water to safely managed water services. This expanded focus to include issues such as water quality requires richer monitoring data and potentially integration of datasets from different sources. Relevant data sets include water point mapping (WPM), the survey of boreholes, wells and other water points, census and household survey data. This study examined inconsistencies between population census and WPM datasets for Cambodia, Liberia and Tanzania, and identified potential barriers to integrating the two datasets to meet monitoring needs. Literatures on numbers of people served per water point were used to convert WPM data to population served by water source type per area and compared with census reports. For Cambodia and Tanzania, discrepancies with census data suggested incomplete WPM coverage. In Liberia, where the data sets were consistent, WPM-derived data on functionality, quantity and quality of drinking water were further combined with census area statistics to generate an enhanced drinking-water access measure for protected wells and springs. The process revealed barriers to integrating census and WPM data, including exclusion of water points not used for drinking by households, matching of census and WPM source types; temporal mismatches between data sources; data quality issues such as missing or implausible data values, and underlying assumptions about population served by different water point technologies. However, integration of these two data sets could be used to identify and rectify gaps in WPM coverage. If WPM databases become more complete and the above barriers are addressed, it could also be used to develop more realistic measures of household drinking-water access for monitoring. Copyright © 2017 Elsevier GmbH. All rights reserved.

  16. The Brainomics/Localizer database.

    Science.gov (United States)

    Papadopoulos Orfanos, Dimitri; Michel, Vincent; Schwartz, Yannick; Pinel, Philippe; Moreno, Antonio; Le Bihan, Denis; Frouin, Vincent

    2017-01-01

    The Brainomics/Localizer database exposes part of the data collected by the in-house Localizer project, which planned to acquire four types of data from volunteer research subjects: anatomical MRI scans, functional MRI data, behavioral and demographic data, and DNA sampling. Over the years, this local project has been collecting such data from hundreds of subjects. We had selected 94 of these subjects for their complete datasets, including all four types of data, as the basis for a prior publication; the Brainomics/Localizer database publishes the data associated with these 94 subjects. Since regulatory rules prevent us from making genetic data available for download, the database serves only anatomical MRI scans, functional MRI data, behavioral and demographic data. To publish this set of heterogeneous data, we use dedicated software based on the open-source CubicWeb semantic web framework. Through genericity in the data model and flexibility in the display of data (web pages, CSV, JSON, XML), CubicWeb helps us expose these complex datasets in original and efficient ways. Copyright © 2015 Elsevier Inc. All rights reserved.

  17. Large Industrial Point Sources in Italy: a focus on mercury concentrations resulting from three seasonal ship-borne measurements

    Directory of Open Access Journals (Sweden)

    Bencardino M.

    2013-04-01

    Full Text Available In Italy there are 25 Large Industrial Point Sources whose mercury emissions in air exceed the established threshold of 10 kg year−1. Many of these mercury point sources, mostly distributed along the Italian coastal area, are located at sites qualified as National Interest Rehabilitation Sites because of documented contamination in qualitative and/or quantitative terms and of potential health impact. Atmospheric mercury emissions related to Italian Large Industrial Point Sources, with a value of 1.04 Mg·yr−1 for 2007, have a not negligible contribution, accounting, on their own, for more than 10% of the total mercury emissions resulting from all activity sectors at a national level. Among others, thermal power stations, pig iron and steel as well as basic inorganic chemical production, result to be the main contributing industrial activities. In order to assess how mercury species concentrations and distribution in the Marine Boundary Layer (MBL change with vicinity to large industrial sites, measurements of atmospheric mercury were performed during three oceanographic campaigns aboard the Research Vessel (R.V. Urania of the Italian CNR. Collection of GEM, GOM and PBM was conducted across the Adriatic sea, during autumn 2004 (27th of October to 12th of November and summer 2005 (17th to 29th of June, and across the Tyrrhenian sea during autumn 2007 (12th of September to 1st October. Analysis were carried out with reference to the period in which the R.V. Urania has stopped close to the main Italian industrial contaminated sites. Explorative statistical parameters of atmospheric mercury species were computed over each single stop-period and then compared with the overall cruise campaign measurements. Results are herein presented and discussed.

  18. Annual seminar on electronic sources of information

    International Nuclear Information System (INIS)

    Ravichandra Rao, I.K.

    2000-03-01

    With the rapid development in IT and the emergence of Internet, a multitude of information sources are now available on electronic media. They include e-journals and other electronic publications - online databases, reference documents, newspapers, magazines, etc. In addition to these online sources, there are thousands of CD-ROM databases. The CD-ROM databases and the online sources are collectively referred to as electronic sources of information. Libraries in no part of the world can afford to ignore these sources. Emergence of these new sources has resulted into a change in the traditional library functions including collection development, acquisitions, cataloguing, user instructions, etc. It is inevitable that in the next five to ten years, special libraries may have to allocate considerable amount towards subscription of e-journals and other e-publications. The papers in this seminar volume discuss several aspects related the theme of the seminar and cover e-journals, different sources available in the Net, classification of electronic sources, online public access catalogues, and different aspects of Internet. Papers relevant to INIS are indexed separately

  19. The XMM-SERVS survey: new XMM-Newton point-source catalog for the XMM-LSS field

    Science.gov (United States)

    Chen, C.-T. J.; Brandt, W. N.; Luo, B.; Ranalli, P.; Yang, G.; Alexander, D. M.; Bauer, F. E.; Kelson, D. D.; Lacy, M.; Nyland, K.; Tozzi, P.; Vito, F.; Cirasuolo, M.; Gilli, R.; Jarvis, M. J.; Lehmer, B. D.; Paolillo, M.; Schneider, D. P.; Shemmer, O.; Smail, I.; Sun, M.; Tanaka, M.; Vaccari, M.; Vignali, C.; Xue, Y. Q.; Banerji, M.; Chow, K. E.; Häußler, B.; Norris, R. P.; Silverman, J. D.; Trump, J. R.

    2018-04-01

    We present an X-ray point-source catalog from the XMM-Large Scale Structure survey region (XMM-LSS), one of the XMM-Spitzer Extragalactic Representative Volume Survey (XMM-SERVS) fields. We target the XMM-LSS region with 1.3 Ms of new XMM-Newton AO-15 observations, transforming the archival X-ray coverage in this region into a 5.3 deg2 contiguous field with uniform X-ray coverage totaling 2.7 Ms of flare-filtered exposure, with a 46 ks median PN exposure time. We provide an X-ray catalog of 5242 sources detected in the soft (0.5-2 keV), hard (2-10 keV), and/or full (0.5-10 keV) bands with a 1% expected spurious fraction determined from simulations. A total of 2381 new X-ray sources are detected compared to previous source catalogs in the same area. Our survey has flux limits of 1.7 × 10-15, 1.3 × 10-14, and 6.5 × 10-15 erg cm-2 s-1 over 90% of its area in the soft, hard, and full bands, respectively, which is comparable to those of the XMM-COSMOS survey. We identify multiwavelength counterpart candidates for 99.9% of the X-ray sources, of which 93% are considered as reliable based on their matching likelihood ratios. The reliabilities of these high-likelihood-ratio counterparts are further confirmed to be ≈97% reliable based on deep Chandra coverage over ≈5% of the XMM-LSS region. Results of multiwavelength identifications are also included in the source catalog, along with basic optical-to-infrared photometry and spectroscopic redshifts from publicly available surveys. We compute photometric redshifts for X-ray sources in 4.5 deg2 of our field where forced-aperture multi-band photometry is available; >70% of the X-ray sources in this subfield have either spectroscopic or high-quality photometric redshifts.

  20. Holocene Sea-Level Database For The Caribbean Region

    Science.gov (United States)

    Khan, N. S.; Horton, B.; Engelhart, S. E.; Peltier, W. R.; Scatena, F. N.; Vane, C. H.; Liu, S.

    2013-12-01

    Holocene relative sea-level (RSL) records from far-field locations are important for understanding the driving mechanisms controlling the nature and timing of the mid-late Holocene reduction in global meltwaters and providing background rates of late Holocene RSL change with which to compare the magnitude of 20th century RSL rise. The Caribbean region has traditionally been considered far-field (i.e., with negligible glacio-isostatic adjustment (GIA) influence), although recent investigations indicate otherwise. Here, we consider the spatial variability in glacio-isostatic, tectonic and local contributions on RSL records from the circum-Caribbean region to infer a Holocene eustatic sea-level signal. We have constructed a database of quality-controlled, spatially comprehensive, Holocene RSL observations for the circum-Caribbean region. The database contains over 500 index points, which locate the position of RSL in time and space. The database incorporates sea-level observations from a latitudinal range of 5°N to 25°N and longitudinal range of 55°W to 90°W. We include sea-level observations from 11 ka BP to present, although the majority of the index points in the database are younger than 8 ka BP. The database is sub-divided into 13 regions based on the distance from the former Laurentide Ice Sheet and regional tectonic setting. The index points were primarily derived from mangrove peat deposits, which in the Caribbean form in the upper half of the tidal range, and corals (predominantly Acropora palmata), the growth of which is constrained to the upper 5 m of water depth. The index points are classified on the basis of their susceptibility to compaction (e.g., intercalated, basal). The influence of temporal changes in tidal range on index points is also considered. The sea-level reconstructions demonstrate that RSL did not exceed the present height (0 m) during the Holocene in the majority of locations, except at sites in Suriname/Guayana and possibly Trinidad

  1. DisFace: A Database of Human Facial Disorders

    Directory of Open Access Journals (Sweden)

    Paramjit Kaur

    2017-10-01

    Full Text Available Face is an integral part of human body by which an individual communicates in the society. Its importance can be highlighted by the fact that a person deprived of face cannot sustain in the living world. In the past few decades, human face has gained attention of several researchers, whether it is related to facial anthropometry, facial disorder, face transplantation or face reconstruction. Several researches have also shown the correlation between neuropsychiatry disorders and human face and also that how face recognition abilities are correlated with these disorders. Currently, several databases exist which contain the facial images of several individuals captured from different sources. The advantage of these databases is that the images in these databases can be used for testing and training purpose. However, in current date no such database exists which would provide not only facial images of individuals; but also the literature concerning the human face, list of several genes controlling human face, list of facial disorders and various tools which work on facial images. Thus, the current research aims at developing a database of human facial disorders using bioinformatics approach. The database will contain information about facial diseases, medications, symptoms, findings, etc. The information will be extracted from several other databases like OMIM, PubChem, Radiopedia, Medline Plus, FDA, etc. and links to them will also be provided. Initially, the diseases specific for human face have been obtained from already created published corpora of literature using text mining approach. Becas tool was used to obtain the specific task.  A dataset will be created and stored in the form of database. It will be a database containing cross-referenced index of human facial diseases, medications, symptoms, signs, etc. Thus, a database on human face with complete existing information about human facial disorders will be developed. The novelty of the

  2. Correction of head movements in positron emission tomography using point source tracking system: a simulation study.

    Science.gov (United States)

    Nazarparvar, Babak; Shamsaei, Mojtaba; Rajabi, Hossein

    2012-01-01

    The motion of the head during brain positron emission tomography (PET) acquisitions has been identified as a source of artifact in the reconstructed image. In this study, a method is described to develop an image-based motion correction technique for correcting the post-acquisition data without using external optical motion-tracking system such as POLARIS. In this technique, GATE has been used to simulate PET brain scan using point sources mounted around the head to accurately monitor the position of the head during the time frames. The measurement of head motion in each frame showed a transformation in the image frame matrix, resulting in a fully corrected data set. Using different kinds of phantoms and motions, the accuracy of the correction method is tested and its applicability to experimental studies is demonstrated as well.

  3. HIPdb: a database of experimentally validated HIV inhibiting peptides.

    Science.gov (United States)

    Qureshi, Abid; Thakur, Nishant; Kumar, Manoj

    2013-01-01

    Besides antiretroviral drugs, peptides have also demonstrated potential to inhibit the Human immunodeficiency virus (HIV). For example, T20 has been discovered to effectively block the HIV entry and was approved by the FDA as a novel anti-HIV peptide (AHP). We have collated all experimental information on AHPs at a single platform. HIPdb is a manually curated database of experimentally verified HIV inhibiting peptides targeting various steps or proteins involved in the life cycle of HIV e.g. fusion, integration, reverse transcription etc. This database provides experimental information of 981 peptides. These are of varying length obtained from natural as well as synthetic sources and tested on different cell lines. Important fields included are peptide sequence, length, source, target, cell line, inhibition/IC(50), assay and reference. The database provides user friendly browse, search, sort and filter options. It also contains useful services like BLAST and 'Map' for alignment with user provided sequences. In addition, predicted structure and physicochemical properties of the peptides are also included. HIPdb database is freely available at http://crdd.osdd.net/servers/hipdb. Comprehensive information of this database will be helpful in selecting/designing effective anti-HIV peptides. Thus it may prove a useful resource to researchers for peptide based therapeutics development.

  4. Harmonization of Databases

    DEFF Research Database (Denmark)

    Charlifue, Susan; Tate, Denise; Biering-Sorensen, Fin

    2016-01-01

    strengths and weaknesses. Efforts to provide a uniform approach to data collection are also reviewed. The databases reviewed offer different approaches to capture important clinical information on SCI. They vary on size, purpose, data points, inclusion of standard outcomes, and technical requirements. Each...... a carefully crafted implementation plan that lists purposes, cost, resources required, and policies to guide such development to establishing a framework for dissemination of data and findings. For the present, taking advantage of the vast but different data already collected over many decades may require...

  5. FCDD: A Database for Fruit Crops Diseases.

    Science.gov (United States)

    Chauhan, Rupal; Jasrai, Yogesh; Pandya, Himanshu; Chaudhari, Suman; Samota, Chand Mal

    2014-01-01

    Fruit Crops Diseases Database (FCDD) requires a number of biotechnology and bioinformatics tools. The FCDD is a unique bioinformatics resource that compiles information about 162 details on fruit crops diseases, diseases type, its causal organism, images, symptoms and their control. The FCDD contains 171 phytochemicals from 25 fruits, their 2D images and their 20 possible sequences. This information has been manually extracted and manually verified from numerous sources, including other electronic databases, textbooks and scientific journals. FCDD is fully searchable and supports extensive text search. The main focus of the FCDD is on providing possible information of fruit crops diseases, which will help in discovery of potential drugs from one of the common bioresource-fruits. The database was developed using MySQL. The database interface is developed in PHP, HTML and JAVA. FCDD is freely available. http://www.fruitcropsdd.com/

  6. Use of point-of-sale data to track usage patterns of residential pesticides: methodology development

    Directory of Open Access Journals (Sweden)

    Chism Bill

    2006-05-01

    Full Text Available Abstract Background Residential-use pesticides have been shown to be a major source of pesticide exposure to people in the United States. However, little is understood about the exposures to household pesticides and the resultant health effects. One reason that little is known about home-use pesticide exposure is the lack of comprehensive data on exposures to pesticides in the home. One method to help ascertain the amount of pesticides present in the home is use of point-of-sale data collected from marketing companies that track product sales to obtain the volume of pesticides sold for home-use. This provides a measure of volume of home-use pesticide. Methods We have constructed a searchable database containing sales data for home-use permethrin-containing pesticides sold by retail stores in the United States from January 1997 through December 2002 in an attempt to develop a tracking method for pesticide. This pilot project was conducted to determine if point-of-sale data would be effective in helping track the purchase of home-use permethrin containing pesticides and if it would stand as a good model for tracking sales of other home-use pesticides. Results There are several limitations associated with this tracking method, including the availability of sales data, market coverage, and geographic resolution. As a result, a fraction of sales data potentially available for reporting is represented in this database. However, the database is sensitive to the number and type of merchants reporting permethrin sales. Further, analysis of the sale of individual products included in the database indicates that year to year variability has a greater impact on reported permethrin sales than the amount sold by each type of merchant. Conclusion We conclude that, while nothing could completely replace a detailed exposure assessment to estimate exposures to home-use pesticides, a point-of-sale database is a useful tool in tracking the purchase of these types

  7. Use of point-of-sale data to track usage patterns of residential pesticides: methodology development.

    Science.gov (United States)

    Bekarian, Nyree; Payne-Sturges, Devon; Edmondson, Stuart; Chism, Bill; Woodruff, Tracey J

    2006-05-25

    Residential-use pesticides have been shown to be a major source of pesticide exposure to people in the United States. However, little is understood about the exposures to household pesticides and the resultant health effects. One reason that little is known about home-use pesticide exposure is the lack of comprehensive data on exposures to pesticides in the home. One method to help ascertain the amount of pesticides present in the home is use of point-of-sale data collected from marketing companies that track product sales to obtain the volume of pesticides sold for home-use. This provides a measure of volume of home-use pesticide. We have constructed a searchable database containing sales data for home-use permethrin-containing pesticides sold by retail stores in the United States from January 1997 through December 2002 in an attempt to develop a tracking method for pesticide. This pilot project was conducted to determine if point-of-sale data would be effective in helping track the purchase of home-use permethrin containing pesticides and if it would stand as a good model for tracking sales of other home-use pesticides. There are several limitations associated with this tracking method, including the availability of sales data, market coverage, and geographic resolution. As a result, a fraction of sales data potentially available for reporting is represented in this database. However, the database is sensitive to the number and type of merchants reporting permethrin sales. Further, analysis of the sale of individual products included in the database indicates that year to year variability has a greater impact on reported permethrin sales than the amount sold by each type of merchant. We conclude that, while nothing could completely replace a detailed exposure assessment to estimate exposures to home-use pesticides, a point-of-sale database is a useful tool in tracking the purchase of these types of pesticides to 1) detect anomalous trends in regional and seasonal

  8. Finger image quality based on singular point localization

    DEFF Research Database (Denmark)

    Wang, Jinghua; Olsen, Martin A.; Busch, Christoph

    2014-01-01

    Singular points are important global features of fingerprints and singular point localization is a crucial step in biometric recognition. Moreover the presence and position of the core point in a captured fingerprint sample can reflect whether the finger is placed properly on the sensor. Therefore...... and analyze the importance of singular points on biometric accuracy. The experiment is based on large scale databases and conducted by relating the measured quality of a fingerprint sample, given by the positions of core points, to the biometric performance. The experimental results show the positions of core...

  9. Experimental properties of gluon and quark jets from a point source

    CERN Document Server

    Abbiendi, G.; Alexander, G.; Allison, John; Altekamp, N.; Anderson, K.J.; Anderson, S.; Arcelli, S.; Asai, S.; Ashby, S.F.; Axen, D.; Azuelos, G.; Ball, A.H.; Barberio, E.; Barlow, Roger J.; Batley, J.R.; Baumann, S.; Bechtluft, J.; Behnke, T.; Bell, Kenneth Watson; Bella, G.; Bellerive, A.; Bentvelsen, S.; Bethke, S.; Betts, S.; Biebel, O.; Biguzzi, A.; Blobel, V.; Bloodworth, I.J.; Bock, P.; Bohme, J.; Bonacorsi, D.; Boutemeur, M.; Braibant, S.; Bright-Thomas, P.; Brigliadori, L.; Brown, Robert M.; Burckhart, H.J.; Capiluppi, P.; Carnegie, R.K.; Carter, A.A.; Carter, J.R.; Chang, C.Y.; Charlton, David G.; Chrisman, D.; Ciocca, C.; Clarke, P.E.L.; Clay, E.; Cohen, I.; Conboy, J.E.; Cooke, O.C.; Couyoumtzelis, C.; Coxe, R.L.; Cuffiani, M.; Dado, S.; Dallavalle, G.Marco; Davis, R.; De Jong, S.; de Roeck, A.; Dervan, P.; Desch, K.; Dienes, B.; Dixit, M.S.; Dubbert, J.; Duchovni, E.; Duckeck, G.; Duerdoth, I.P.; Estabrooks, P.G.; Etzion, E.; Fabbri, F.; Fanfani, A.; Fanti, M.; Faust, A.A.; Fiedler, F.; Fierro, M.; Fleck, I.; Folman, R.; Frey, A.; Furtjes, A.; Futyan, D.I.; Gagnon, P.; Gary, J.W.; Gascon, J.; Gascon-Shotkin, S.M.; Gaycken, G.; Geich-Gimbel, C.; Giacomelli, G.; Giacomelli, P.; Gibson, V.; Gibson, W.R.; Gingrich, D.M.; Glenzinski, D.; Goldberg, J.; Gorn, W.; Grandi, C.; Graham, K.; Gross, E.; Grunhaus, J.; Gruwe, M.; Hanson, G.G.; Hansroul, M.; Hapke, M.; Harder, K.; Harel, A.; Hargrove, C.K.; Hauschild, M.; Hawkes, C.M.; Hawkings, R.; Hemingway, R.J.; Herndon, M.; Herten, G.; Heuer, R.D.; Hildreth, M.D.; Hill, J.C.; Hobson, P.R.; Hoch, M.; Hocker, James Andrew; Hoffman, Kara Dion; Homer, R.J.; Honma, A.K.; Horvath, D.; Hossain, K.R.; Howard, R.; Huntemeyer, P.; Igo-Kemenes, P.; Imrie, D.C.; Ishii, K.; Jacob, F.R.; Jawahery, A.; Jeremie, H.; Jimack, M.; Jones, C.R.; Jovanovic, P.; Junk, T.R.; Kanzaki, J.; Karlen, D.; Kartvelishvili, V.; Kawagoe, K.; Kawamoto, T.; Kayal, P.I.; Keeler, R.K.; Kellogg, R.G.; Kennedy, B.W.; Kim, D.H.; Klier, A.; Kobayashi, T.; Kobel, M.; Kokott, T.P.; Kolrep, M.; Komamiya, S.; Kowalewski, Robert V.; Kress, T.; Krieger, P.; von Krogh, J.; Kuhl, T.; Kyberd, P.; Lafferty, G.D.; Landsman, H.; Lanske, D.; Lauber, J.; Lautenschlager, S.R.; Lawson, I.; Layter, J.G.; Lee, A.M.; Lellouch, D.; Letts, J.; Levinson, L.; Liebisch, R.; List, B.; Littlewood, C.; Lloyd, A.W.; Lloyd, S.L.; Loebinger, F.K.; Long, G.D.; Losty, M.J.; Lu, J.; Ludwig, J.; Lui, D.; Macchiolo, A.; Macpherson, A.; Mader, W.; Mannelli, M.; Marcellini, S.; Markopoulos, C.; Martin, A.J.; Martin, J.P.; Martinez, G.; Mashimo, T.; Mattig, Peter; McDonald, W.John; McKenna, J.; Mckigney, E.A.; McMahon, T.J.; McPherson, R.A.; Meijers, F.; Menke, S.; Merritt, F.S.; Mes, H.; Meyer, J.; Michelini, A.; Mihara, S.; Mikenberg, G.; Miller, D.J.; Mir, R.; Mohr, W.; Montanari, A.; Mori, T.; Nagai, K.; Nakamura, I.; Neal, H.A.; Nisius, R.; O'Neale, S.W.; Oakham, F.G.; Odorici, F.; Ogren, H.O.; Oreglia, M.J.; Orito, S.; Palinkas, J.; Pasztor, G.; Pater, J.R.; Patrick, G.N.; Patt, J.; Perez-Ochoa, R.; Petzold, S.; Pfeifenschneider, P.; Pilcher, J.E.; Pinfold, J.; Plane, David E.; Poffenberger, P.; Poli, B.; Polok, J.; Przybycien, M.; Rembser, C.; Rick, H.; Robertson, S.; Robins, S.A.; Rodning, N.; Roney, J.M.; Rosati, S.; Roscoe, K.; Rossi, A.M.; Rozen, Y.; Runge, K.; Runolfsson, O.; Rust, D.R.; Sachs, K.; Saeki, T.; Sahr, O.; Sang, W.M.; Sarkisian, E.K.G.; Sbarra, C.; Schaile, A.D.; Schaile, O.; Scharff-Hansen, P.; Schieck, J.; Schmitt, S.; Schoning, A.; Schroder, Matthias; Schumacher, M.; Schwick, C.; Scott, W.G.; Seuster, R.; Shears, T.G.; Shen, B.C.; Shepherd-Themistocleous, C.H.; Sherwood, P.; Siroli, G.P.; Sittler, A.; Skuja, A.; Smith, A.M.; Snow, G.A.; Sobie, R.; Soldner-Rembold, S.; Spagnolo, S.; Sproston, M.; Stahl, A.; Stephens, K.; Steuerer, J.; Stoll, K.; Strom, David M.; Strohmer, R.; Surrow, B.; Talbot, S.D.; Taras, P.; Tarem, S.; Teuscher, R.; Thiergen, M.; Thomas, J.; Thomson, M.A.; Torrence, E.; Towers, S.; Trigger, I.; Trocsanyi, Z.; Tsur, E.; Turcot, A.S.; Turner-Watson, M.F.; Ueda, I.; Van Kooten, Rick J.; Vannerem, P.; Verzocchi, M.; Voss, H.; Wackerle, F.; Wagner, A.; Ward, C.P.; Ward, D.R.; Watkins, P.M.; Watson, A.T.; Watson, N.K.; Wells, P.S.; Wermes, N.; White, J.S.; Wilson, G.W.; Wilson, J.A.; Wyatt, T.R.; Yamashita, S.; Yekutieli, G.; Zacek, V.; Zer-Zion, D.

    1999-01-01

    Gluon jets are identified in hadronic Z0 decays as all the particles in a hemisphere opposite to a hemisphere containing two tagged quark jets. Gluon jets defined in this manner are equivalent to gluon jets produced from a color singlet point source and thus correspond to the definition employed for most theoretical calculations. In a separate stage of the analysis, we select quark jets in a manner to correspond to calculations, as the particles in hemispheres of flavor tagged light quark (uds) events. We present the distributions of rapidity, scaled energy, the logarithm of the momentum, and transverse momentum with respect to the jet axes, for charged particles in these gluon and quark jets. We also examine the charged particle multiplicity distributions of the jets in restricted intervals of rapidity. For soft particles at large transverse momentum, we observe the charged particle multiplicity ratio of gluon to quark jets to be 2.29 +- 0.09 +- 0.15 in agreement with the prediction that this ratio should ap...

  10. CHID: a unique health information and education database.

    OpenAIRE

    Lunin, L F; Stein, R S

    1987-01-01

    The public's growing interest in health information and the health professions' increasing need to locate health education materials can be answered in part by the new Combined Health Information Database (CHID). This unique database focuses on materials and programs in professional and patient education, general health education, and community risk reduction. Accessible through BRS, CHID suggests sources for procuring brochures, pamphlets, articles, and films on community services, programs ...

  11. An Intelligent Assistant for Construction of Terrain Databases

    OpenAIRE

    Rowe, Neil C.; Reed, Chris; Jackson, Leroy; Baer, Wolfgang

    1998-01-01

    1998 Command and Control Research and Technology Symposium, Monterey CA, June 1998, 481-486. We describe TELLUSPLAN, an intelligent assistant for the problem of bargaining between user goals and system resources in the integration of terrain databases from separate source databases. TELLUSPLAN uses nondeterministic methods from artificial intelligence and a detailed cost model to infer the most reasonable compromise with the user's needs. Supported by the Army Artificial Int...

  12. SINBAD: Shielding integral benchmark archive and database

    International Nuclear Information System (INIS)

    Hunter, H.T.; Ingersoll, D.T.; Roussin, R.W.

    1996-01-01

    SINBAD is a new electronic database developed to store a variety of radiation shielding benchmark data so that users can easily retrieve and incorporate the data into their calculations. SINBAD is an excellent data source for users who require the quality assurance necessary in developing cross-section libraries or radiation transport codes. The future needs of the scientific community are best served by the electronic database format of SINBAD and its user-friendly interface, combined with its data accuracy and integrity

  13. The IAEA's Illicit Trafficking Database Programme

    International Nuclear Information System (INIS)

    Anzelon, G.; Hammond, W.; Nicholas, M.

    2001-01-01

    Full text: As part of its overall programme on nuclear material security, the IAEA has since 1995 maintained a database of incidents of trafficking in nuclear materials and other radioactive sources. The Illicit Trafficking Database Programme (ITDP) is intended to assist Member States by alerting them to current incidents, by facilitating exchange of reliable, detailed information about incidents, and by identifying any common threads or trends that might assist States in combating illicit trafficking. The ITDP also seeks to better inform the public by providing basic information to the media concerning illicit trafficking events. Approximately 70 States have joined this programme for collecting and sharing information on trafficking incidents. Reporting States have the opportunity to designate what information may be shared with other States and what may be shared with the public. In cases where the IAEA's first information about a possible incident comes from news media or other open sources rather than from a State notification, the information first is evaluated, and then, if warranted, the relevant State or States are contacted to request confirmation or clarification of an alleged incident. During 2000, as a result of experience gained working with information on illicit nuclear trafficking, the IAEA developed of a flexible and comprehensive new database system. The new system has an open architecture that accommodates structured information from States, in-house information, open-source articles, and other information sources, such as pictures, maps and web links. The graphical user interface allows data entry, maintenance and standard and ad-hoc reporting. The system also is linked to a Web-based query engine, which enables searching of both structured and open-source information. For the period 1 January 1993 through 31 March 2001, the database recorded more than 550 incidents, of which about two-thirds have been confirmed by States. (Some of these

  14. Managing expectations: assessment of chemistry databases generated by automated extraction of chemical structures from patents.

    Science.gov (United States)

    Senger, Stefan; Bartek, Luca; Papadatos, George; Gaulton, Anna

    2015-12-01

    First public disclosure of new chemical entities often takes place in patents, which makes them an important source of information. However, with an ever increasing number of patent applications, manual processing and curation on such a large scale becomes even more challenging. An alternative approach better suited for this large corpus of documents is the automated extraction of chemical structures. A number of patent chemistry databases generated by using the latter approach are now available but little is known that can help to manage expectations when using them. This study aims to address this by comparing two such freely available sources, SureChEMBL and IBM SIIP (IBM Strategic Intellectual Property Insight Platform), with manually curated commercial databases. When looking at the percentage of chemical structures successfully extracted from a set of patents, using SciFinder as our reference, 59 and 51 % were also found in our comparison in SureChEMBL and IBM SIIP, respectively. When performing this comparison with compounds as starting point, i.e. establishing if for a list of compounds the databases provide the links between chemical structures and patents they appear in, we obtained similar results. SureChEMBL and IBM SIIP found 62 and 59 %, respectively, of the compound-patent pairs obtained from Reaxys. In our comparison of automatically generated vs. manually curated patent chemistry databases, the former successfully provided approximately 60 % of links between chemical structure and patents. It needs to be stressed that only a very limited number of patents and compound-patent pairs were used for our comparison. Nevertheless, our results will hopefully help to manage expectations of users of patent chemistry databases of this type and provide a useful framework for more studies like ours as well as guide future developments of the workflows used for the automated extraction of chemical structures from patents. The challenges we have encountered

  15. Some Aspects of Process Computers Configuration Control in Nuclear Power Plant Krsko - Process Computer Signal Configuration Database (PCSCDB)

    International Nuclear Information System (INIS)

    Mandic, D.; Kocnar, R.; Sucic, B.

    2002-01-01

    During the operation of NEK and other nuclear power plants it has been recognized that certain issues related to the usage of digital equipment and associated software in NPP technological process protection, control and monitoring, is not adequately addressed in the existing programs and procedures. The term and the process of Process Computers Configuration Control joins three 10CFR50 Appendix B quality requirements of Process Computers application in NPP: Design Control, Document Control and Identification and Control of Materials, Parts and Components. This paper describes Process Computer Signal Configuration Database (PCSCDB), that was developed and implemented in order to resolve some aspects of Process Computer Configuration Control related to the signals or database points that exist in the life cycle of different Process Computer Systems (PCS) in Nuclear Power Plant Krsko. PCSCDB is controlled, master database, related to the definition and description of the configurable database points associated with all Process Computer Systems in NEK. PCSCDB holds attributes related to the configuration of addressable and configurable real time database points and attributes related to the signal life cycle references and history data such as: Input/Output signals, Manually Input database points, Program constants, Setpoints, Calculated (by application program or SCADA calculation tools) database points, Control Flags (example: enable / disable certain program feature) Signal acquisition design references to the DCM (Document Control Module Application software for document control within Management Information System - MIS) and MECL (Master Equipment and Component List MIS Application software for identification and configuration control of plant equipment and components) Usage of particular database point in particular application software packages, and in the man-machine interface features (display mimics, printout reports, ...) Signals history (EEAR Engineering

  16. ARTI refrigerant database

    Energy Technology Data Exchange (ETDEWEB)

    Calm, J.M. [Calm (James M.), Great Falls, VA (United States)

    1998-08-01

    The Refrigerant Database is an information system on alternative refrigerants, associated lubricants, and their use in air conditioning and refrigeration. It consolidates and facilitates access to property, compatibility, environmental, safety, application and other information. It provides corresponding information on older refrigerants, to assist manufactures and those using alternative refrigerants, to make comparisons and determine differences. The underlying purpose is to accelerate phase out of chemical compounds of environmental concern. The database provides bibliographic citations and abstracts for publications that may be useful in research and design of air-conditioning and refrigeration equipment. The complete documents are not included, though some may be added at a later date. The database identifies sources of specific information on many refrigerants including propane, ammonia, water, carbon dioxide, propylene, ethers, and others as well as azeotropic and zeotropic blends of these fluids. It addresses lubricants including alkylbenzene, polyalkylene glycol, polyolester, and other synthetics as well as mineral oils. It also references documents addressing compatibility of refrigerants and lubricants with metals, plastics, elastomers, motor insulation, and other materials used in refrigerant circuits. Incomplete citations or abstracts are provided for some documents. They are included to accelerate availability of the information and will be completed or replaced in future updates.

  17. Intra-disciplinary differences in database coverage and the consequences for bibliometric research

    DEFF Research Database (Denmark)

    Faber Frandsen, Tove; Nicolaisen, Jeppe

    2008-01-01

    Bibliographic databases (including databases based on open access) are routinely used for bibliometric research. The value of a specific database depends to a large extent on the coverage of the discipline(s) under study. A number of studies have determined the coverage of databases in specific d...... and psychology). The point extends to include both the uneven coverage of specialties and research traditions. The implications for bibliometric research are discussed, and precautions which need to be taken are outlined. ...

  18. Cross-correlation of the cosmic microwave background with the 2MASS galaxy survey: Signatures of dark energy, hot gas, and point sources

    International Nuclear Information System (INIS)

    Afshordi, Niayesh; Loh, Yeong-Shang; Strauss, Michael A.

    2004-01-01

    We cross-correlate the cosmic microwave background temperature anisotropies observed by the Wilkinson Microwave Anisotropy Probe (WMAP) with the projected distribution of extended sources in the Two Micron All Sky Survey (2MASS). By modeling the theoretical expectation for this signal, we extract the signatures of dark energy [integrated Sachs-Wolfe effect (ISW)], hot gas [thermal Sunyaev-Zeldovich (SZ) effect], and microwave point sources in the cross-correlation. Our strongest signal is the thermal SZ, at the 3.1-3.7σ level, which is consistent with the theoretical prediction based on observations of x-ray clusters. We also see the ISW signal at the 2.5σ level, which is consistent with the expected value for the concordance ΛCDM cosmology, and is an independent signature of the presence of dark energy in the Universe. Finally, we see the signature of microwave point sources at the 2.7σ level

  19. Database Description - Yeast Interacting Proteins Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Yeast Interacting Proteins Database Database Description General information of database Database... name Yeast Interacting Proteins Database Alternative name - DOI 10.18908/lsdba.nbdc00742-000 Creator C...-ken 277-8561 Tel: +81-4-7136-3989 FAX: +81-4-7136-3979 E-mail : Database classif...s cerevisiae Taxonomy ID: 4932 Database description Information on interactions and related information obta...l Acad Sci U S A. 2001 Apr 10;98(8):4569-74. Epub 2001 Mar 13. External Links: Original website information Database

  20. Healthcare databases in Europe for studying medicine use and safety during pregnancy

    DEFF Research Database (Denmark)

    Charlton, Rachel A; Neville, Amanda J; Jordan, Sue

    2014-01-01

    data recorded by primary-care practitioners. All databases captured maternal co-prescribing and a measure of socioeconomic status. CONCLUSION: This study suggests that within Europe, electronic healthcare databases may be valuable sources of data for evaluating medicine use and safety during pregnancy......PURPOSE: The aim of this study was to describe a number of electronic healthcare databases in Europe in terms of the population covered, the source of the data captured and the availability of data on key variables required for evaluating medicine use and medicine safety during pregnancy. METHODS....... The suitability of a particular database, however, will depend on the research question, the type of medicine to be evaluated, the prevalence of its use and any adverse outcomes of interest. © 2014 The Authors. Pharmacoepidemiology and Drug Safety published by John Wiley & Sons, Ltd....

  1. THE EXTRAGALACTIC DISTANCE DATABASE

    International Nuclear Information System (INIS)

    Tully, R. Brent; Courtois, Helene M.; Jacobs, Bradley A.; Rizzi, Luca; Shaya, Edward J.; Makarov, Dmitry I.

    2009-01-01

    A database can be accessed on the Web at http://edd.ifa.hawaii.edu that was developed to promote access to information related to galaxy distances. The database has three functional components. First, tables from many literature sources have been gathered and enhanced with links through a distinct galaxy naming convention. Second, comparisons of results both at the levels of parameters and of techniques have begun and are continuing, leading to increasing homogeneity and consistency of distance measurements. Third, new material is presented arising from ongoing observational programs at the University of Hawaii 2.2 m telescope, radio telescopes at Green Bank, Arecibo, and Parkes and with the Hubble Space Telescope. This new observational material is made available in tandem with related material drawn from archives and passed through common analysis pipelines.

  2. Non-point Source Pollution Modeling Using Geographic Information System (GIS for Representing Best Management Practices (BMP in the Gorganrood Watershed

    Directory of Open Access Journals (Sweden)

    Z. Pasandidehfard

    2014-09-01

    Full Text Available The most important pollutants that cause water pollution are nitrogen and phosphorus from agricultural runoff called Non-Point Source Pollution (NPS. To solve this problem, management practices known as BMPs or Best Management Practices are applied. One of the common methods for Non-Point Source Pollution prediction is modeling. By modeling, efficiency of many practices can be tested before application. In this study, land use changes were studied from the years 1984 till 2010 that showed an increase in agricultural lands from 516908.52 to 630737.19 ha and expansion of cities from 5237.87 to 15487.59 ha and roads from 9666.07 to 11430.24 ha. Using L-THIA model (from nonpoint source pollution models for both land use categories, the amount of pollutant and the volume of runoff were calculated that showed high growth. Then, the seventh sub-basin was recognized as a critical zone in terms of pollution among the sub-basins. In the end, land use change was considered as a BMP using Multi-Criteria Evaluation (MCE based on which a more suitable land use map was produced. After producing the new land use map, L-THIA model was run again and the result of the model was compared to the actual land use to show the effect of this BMP. Runoff volume decreased from 367.5 to 308.6 M3/ha and nitrogen in runoff was reduced from 3.26 to 1.58 mg/L and water BOD from 3.61 to 2.13 mg/L. Other pollutants also showed high reduction. In the end, land use change is confirmed as an effective BMP for Non-Point Source Pollution reduction.

  3. Interactive bibliographical database on color

    Science.gov (United States)

    Caivano, Jose L.

    2002-06-01

    The paper describes the methodology and results of a project under development, aimed at the elaboration of an interactive bibliographical database on color in all fields of application: philosophy, psychology, semiotics, education, anthropology, physical and natural sciences, biology, medicine, technology, industry, architecture and design, arts, linguistics, geography, history. The project is initially based upon an already developed bibliography, published in different journals, updated in various opportunities, and now available at the Internet, with more than 2,000 entries. The interactive database will amplify that bibliography, incorporating hyperlinks and contents (indexes, abstracts, keywords, introductions, or eventually the complete document), and devising mechanisms for information retrieval. The sources to be included are: books, doctoral dissertations, multimedia publications, reference works. The main arrangement will be chronological, but the design of the database will allow rearrangements or selections by different fields: subject, Decimal Classification System, author, language, country, publisher, etc. A further project is to develop another database, including color-specialized journals or newsletters, and articles on color published in international journals, arranged in this case by journal name and date of publication, but allowing also rearrangements or selections by author, subject and keywords.

  4. The Chandra Source Catalog: Storage and Interfaces

    Science.gov (United States)

    van Stone, David; Harbo, Peter N.; Tibbetts, Michael S.; Zografou, Panagoula; Evans, Ian N.; Primini, Francis A.; Glotfelty, Kenny J.; Anderson, Craig S.; Bonaventura, Nina R.; Chen, Judy C.; Davis, John E.; Doe, Stephen M.; Evans, Janet D.; Fabbiano, Giuseppina; Galle, Elizabeth C.; Gibbs, Danny G., II; Grier, John D.; Hain, Roger; Hall, Diane M.; He, Xiang Qun (Helen); Houck, John C.; Karovska, Margarita; Kashyap, Vinay L.; Lauer, Jennifer; McCollough, Michael L.; McDowell, Jonathan C.; Miller, Joseph B.; Mitschang, Arik W.; Morgan, Douglas L.; Mossman, Amy E.; Nichols, Joy S.; Nowak, Michael A.; Plummer, David A.; Refsdal, Brian L.; Rots, Arnold H.; Siemiginowska, Aneta L.; Sundheim, Beth A.; Winkelman, Sherry L.

    2009-09-01

    The Chandra Source Catalog (CSC) is part of the Chandra Data Archive (CDA) at the Chandra X-ray Center. The catalog contains source properties and associated data objects such as images, spectra, and lightcurves. The source properties are stored in relational databases and the data objects are stored in files with their metadata stored in databases. The CDA supports different versions of the catalog: multiple fixed release versions and a live database version. There are several interfaces to the catalog: CSCview, a graphical interface for building and submitting queries and for retrieving data objects; a command-line interface for property and source searches using ADQL; and VO-compliant services discoverable though the VO registry. This poster describes the structure of the catalog and provides an overview of the interfaces.

  5. Update History of This Database - Trypanosomes Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Trypanosomes Database Update History of This Database Date Update contents 2014/05/07 The co...ntact information is corrected. The features and manner of utilization of the database are corrected. 2014/02/04 Trypanosomes Databas...e English archive site is opened. 2011/04/04 Trypanosomes Database ( http://www.tan...paku.org/tdb/ ) is opened. About This Database Database Description Download Lice...nse Update History of This Database Site Policy | Contact Us Update History of This Database - Trypanosomes Database | LSDB Archive ...

  6. Effects of pointing compared with naming and observing during encoding on item and source memory in young and older adults

    NARCIS (Netherlands)

    Ouwehand, Kim; Gog, Tamara van; Paas, Fred

    2016-01-01

    Research showed that source memory functioning declines with ageing. Evidence suggests that encoding visual stimuli with manual pointing in addition to visual observation can have a positive effect on spatial memory compared with visual observation only. The present study investigated whether

  7. Field-scale operation of methane biofiltration systems to mitigate point source methane emissions

    International Nuclear Information System (INIS)

    Hettiarachchi, Vijayamala C.; Hettiaratchi, Patrick J.; Mehrotra, Anil K.; Kumar, Sunil

    2011-01-01

    Methane biofiltration (MBF) is a novel low-cost technique for reducing low volume point source emissions of methane (CH 4 ). MBF uses a granular medium, such as soil or compost, to support the growth of methanotrophic bacteria responsible for converting CH 4 to carbon dioxide (CO 2 ) and water (H 2 O). A field research program was undertaken to evaluate the potential to treat low volume point source engineered CH 4 emissions using an MBF at a natural gas monitoring station. A new comprehensive three-dimensional numerical model was developed incorporating advection-diffusive flow of gas, biological reactions and heat and moisture flow. The one-dimensional version of this model was used as a guiding tool for designing and operating the MBF. The long-term monitoring results of the field MBF are also presented. The field MBF operated with no control of precipitation, evaporation, and temperature, provided more than 80% of CH 4 oxidation throughout spring, summer, and fall seasons. The numerical model was able to predict the CH 4 oxidation behavior of the field MBF with high accuracy. The numerical model simulations are presented for estimating CH 4 oxidation efficiencies under various operating conditions, including different filter bed depths and CH 4 flux rates. The field observations as well as numerical model simulations indicated that the long-term performance of MBFs is strongly dependent on environmental factors, such as ambient temperature and precipitation. - Highlights: → One-dimensional version of the model was used as a guiding tool for designing and operating the MBF. → Mathematical model predicted CH 4 oxidation behaviors of the field MBF with high accuracy i.e. (> 80 %). → Performance of MBF is dependent on ambient temperature and precipitation. - The developed numerical model simulations and field observations for estimating CH 4 oxidation efficiencies under various operating conditions indicate that the long-term performance of MBFs is strongly

  8. Summary of earthquake experience database

    International Nuclear Information System (INIS)

    1999-01-01

    Strong-motion earthquakes frequently occur throughout the Pacific Basin, where power plants or industrial facilities are included in the affected areas. By studying the performance of these earthquake-affected (or database) facilities, a large inventory of various types of equipment installations can be compiled that have experienced substantial seismic motion. The primary purposes of the seismic experience database are summarized as follows: to determine the most common sources of seismic damage, or adverse effects, on equipment installations typical of industrial facilities; to determine the thresholds of seismic motion corresponding to various types of seismic damage; to determine the general performance of equipment during earthquakes, regardless of the levels of seismic motion; to determine minimum standards in equipment construction and installation, based on past experience, to assure the ability to withstand anticipated seismic loads. To summarize, the primary assumption in compiling an experience database is that the actual seismic hazard to industrial installations is best demonstrated by the performance of similar installations in past earthquakes

  9. The STRING database in 2017

    DEFF Research Database (Denmark)

    Szklarczyk, Damian; Morris, John H; Cook, Helen

    2017-01-01

    A system-wide understanding of cellular function requires knowledge of all functional interactions between the expressed proteins. The STRING database aims to collect and integrate this information, by consolidating known and predicted protein-protein association data for a large number of organi......A system-wide understanding of cellular function requires knowledge of all functional interactions between the expressed proteins. The STRING database aims to collect and integrate this information, by consolidating known and predicted protein-protein association data for a large number...... of organisms. The associations in STRING include direct (physical) interactions, as well as indirect (functional) interactions, as long as both are specific and biologically meaningful. Apart from collecting and reassessing available experimental data on protein-protein interactions, and importing known...... pathways and protein complexes from curated databases, interaction predictions are derived from the following sources: (i) systematic co-expression analysis, (ii) detection of shared selective signals across genomes, (iii) automated text-mining of the scientific literature and (iv) computational transfer...

  10. Atmospheric observations for quantifying emissions of point-source synthetic greenhouse gases (CF4, NF3 and HFC-23)

    Science.gov (United States)

    Arnold, Tim; Manning, Alistair J.; Li, Shanlan; Kim, Jooil; Park, Sunyoung; Fraser, Paul J.; Mitrevski, Blagoj; Steele, L. Paul; Krummel, Paul B.; Mühle, Jens; Weiss, Ray F.

    2016-04-01

    The fluorinated species carbon tetrafluoride (CF4; PFC-14), nitrogen trifluoride (NF3) and trifluoromethane (CHF3; HFC-23) are potent greenhouse gases with 100-year global warming potentials of 6,630, 16,100 and 12,400, respectively. Unlike the majority of CFC-replacement compounds that are emitted from fugitive and mobile emission sources, these gases are largely emitted from large single point sources - semiconductor manufacturing facilities (all three), aluminium smelting plants (CF4) and chlorodifluoromethane factories (HFC-23). In this work we show the potential for atmospheric measurements to understand regional sources of these gases and to highlight emission 'hotspots'. We target our analysis on measurements from two Advanced Global Atmospheric Gases Experiment (AGAGE) long term monitoring sites that are particularly sensitive to regional emissions of these gases: Gosan on Jeju Island in the Republic of Korea and Cape Grim on Tasmania in Australia. These sites measure CF4, NF3 and HFC-23 alongside a suite of greenhouse and stratospheric ozone depleting gases every two hours using automated in situ gas-chromatography mass-spectrometry instrumentation. We couple each measurement to an analysis of air history using the regional atmospheric transport model NAME (Numerical Atmospheric dispersion Modelling Environment) driven by 3D meteorology from the Met Office's Unified Model, and use a Bayesian inverse method (InTEM - Inversion Technique for Emission Modelling) to calculate yearly emission changes over a decade (2005-2015) at high spatial resolution. At present these gases make a small contribution to global radiative forcing, however, given that their impact could rise significantly and that point sources of such gases can be mitigated, atmospheric monitoring could be an important tool for aiding emissions reduction policy.

  11. Advanced approaches to intelligent information and database systems

    CERN Document Server

    Boonjing, Veera; Chittayasothorn, Suphamit

    2014-01-01

    This book consists of 35 chapters presenting different theoretical and practical aspects of Intelligent Information and Database Systems. Nowadays both Intelligent and Database Systems are applied in most of the areas of human activities which necessitates further research in these areas. In this book various interesting issues related to the intelligent information models and methods as well as their advanced applications, database systems applications, data models and their analysis, and digital multimedia methods and applications are presented and discussed both from the practical and theoretical points of view. The book is organized in four parts devoted to intelligent systems models and methods, intelligent systems advanced applications, database systems methods and applications, and multimedia systems methods and applications. The book will be interesting for both practitioners and researchers, especially graduate and PhD students of information technology and computer science, as well more experienced ...

  12. IMPLEMENTATION OF COLUMN-ORIENTED DATABASE IN POSTGRESQL FOR OPTIMIZATION OF READ-ONLY QUERIES

    OpenAIRE

    Aditi D. Andurkar

    2012-01-01

    The era of column-oriented database systems has truly begun with open source database systems like C-Store, MonetDb, LucidDb and commercial ones like Vertica. Column-oriented database stores data column-by-column which means it stores information of single attribute collectively. The need for Column-oriented database arose from the need of business intelligence for efficient decision making where traditional row-oriented database gives poor performance. PostgreSql is an open so...

  13. Point-source and diffuse high-energy neutrino emission from Type IIn supernovae

    Science.gov (United States)

    Petropoulou, M.; Coenders, S.; Vasilopoulos, G.; Kamble, A.; Sironi, L.

    2017-09-01

    Type IIn supernovae (SNe), a rare subclass of core collapse SNe, explode in dense circumstellar media that have been modified by the SNe progenitors at their last evolutionary stages. The interaction of the freely expanding SN ejecta with the circumstellar medium gives rise to a shock wave propagating in the dense SN environment, which may accelerate protons to multi-PeV energies. Inelastic proton-proton collisions between the shock-accelerated protons and those of the circumstellar medium lead to multimessenger signatures. Here, we evaluate the possible neutrino signal of Type IIn SNe and compare with IceCube observations. We employ a Monte Carlo method for the calculation of the diffuse neutrino emission from the SN IIn class to account for the spread in their properties. The cumulative neutrino emission is found to be ˜10 per cent of the observed IceCube neutrino flux above 60 TeV. Type IIn SNe would be the dominant component of the diffuse astrophysical flux, only if 4 per cent of all core collapse SNe were of this type and 20-30 per cent of the shock energy was channeled to accelerated protons. Lower values of the acceleration efficiency are accessible by the observation of a single Type IIn SN as a neutrino point source with IceCube using up-going muon neutrinos. Such an identification is possible in the first year following the SN shock breakout for sources within 20 Mpc.

  14. Improvement of correlation-based centroiding methods for point source Shack-Hartmann wavefront sensor

    Science.gov (United States)

    Li, Xuxu; Li, Xinyang; wang, Caixia

    2018-03-01

    This paper proposes an efficient approach to decrease the computational costs of correlation-based centroiding methods used for point source Shack-Hartmann wavefront sensors. Four typical similarity functions have been compared, i.e. the absolute difference function (ADF), ADF square (ADF2), square difference function (SDF), and cross-correlation function (CCF) using the Gaussian spot model. By combining them with fast search algorithms, such as three-step search (TSS), two-dimensional logarithmic search (TDL), cross search (CS), and orthogonal search (OS), computational costs can be reduced drastically without affecting the accuracy of centroid detection. Specifically, OS reduces calculation consumption by 90%. A comprehensive simulation indicates that CCF exhibits a better performance than other functions under various light-level conditions. Besides, the effectiveness of fast search algorithms has been verified.

  15. Proposal for a high-energy nuclear database

    International Nuclear Information System (INIS)

    Brown, D.A.; Vogt, R.

    2006-01-01

    We propose to develop a high-energy heavy-ion experimental database and make it accessible to the scientific community through an on-line interface. This database will be searchable and cross-indexed with relevant publications, including published detector descriptions. Since this database will be a community resource, it requires the high-energy nuclear physics community's financial and manpower support. This database should eventually contain all published data from Bevalac, AGS and SPS to RHIC and LHC energies, proton-proton to nucleus-nucleus collisions as well as other relevant systems, and all measured observables. Such a database would have tremendous scientific payoff as it makes systematic studies easier and allows simpler benchmarking of theoretical models to a broad range of old and new experiments. Furthermore, there is a growing need for compilations of high-energy nuclear data for applications including stockpile stewardship, technology development for inertial confinement fusion and target and source development for upcoming facilities such as the Next Linear Collider. To enhance the utility of this database, we propose periodically performing evaluations of the data and summarizing the results in topical reviews. (author)

  16. Proposal for a High Energy Nuclear Database

    International Nuclear Information System (INIS)

    Brown, D A; Vogt, R

    2005-01-01

    The authors propose to develop a high-energy heavy-ion experimental database and make it accessible to the scientific community through an on-line interface. This database will be searchable and cross-indexed with relevant publications, including published detector descriptions. Since this database will be a community resource, it requires the high-energy nuclear physics community's financial and manpower support. This database should eventually contain all published data from Bevalac, AGS and SPS to RHIC and CERN-LHC energies, proton-proton to nucleus-nucleus collisions as well as other relevant systems, and all measured observables. Such a database would have tremendous scientific payoff as it makes systematic studies easier and allows simpler benchmarking of theoretical models to a broad range of old and new experiments. Furthermore, there is a growing need for compilations of high-energy nuclear data for applications including stockpile stewardship, technology development for inertial confinement fusion and target and source development for upcoming facilities such as the Next Linear Collider. To enhance the utility of this database, they propose periodically performing evaluations of the data and summarizing the results in topical reviews

  17. [Method of traditional Chinese medicine formula design based on 3D-database pharmacophore search and patent retrieval].

    Science.gov (United States)

    He, Yu-su; Sun, Zhi-yi; Zhang, Yan-ling

    2014-11-01

    By using the pharmacophore model of mineralocorticoid receptor antagonists as a starting point, the experiment stud- ies the method of traditional Chinese medicine formula design for anti-hypertensive. Pharmacophore models were generated by 3D-QSAR pharmacophore (Hypogen) program of the DS3.5, based on the training set composed of 33 mineralocorticoid receptor antagonists. The best pharmacophore model consisted of two Hydrogen-bond acceptors, three Hydrophobic and four excluded volumes. Its correlation coefficient of training set and test set, N, and CAI value were 0.9534, 0.6748, 2.878, and 1.119. According to the database screening, 1700 active compounds from 86 source plant were obtained. Because of lacking of available anti-hypertensive medi cation strategy in traditional theory, this article takes advantage of patent retrieval in world traditional medicine patent database, in order to design drug formula. Finally, two formulae was obtained for antihypertensive.

  18. Molecule database framework: a framework for creating database applications with chemical structure search capability.

    Science.gov (United States)

    Kiener, Joos

    2013-12-11

    Research in organic chemistry generates samples of novel chemicals together with their properties and other related data. The involved scientists must be able to store this data and search it by chemical structure. There are commercial solutions for common needs like chemical registration systems or electronic lab notebooks. However for specific requirements of in-house databases and processes no such solutions exist. Another issue is that commercial solutions have the risk of vendor lock-in and may require an expensive license of a proprietary relational database management system. To speed up and simplify the development for applications that require chemical structure search capabilities, I have developed Molecule Database Framework. The framework abstracts the storing and searching of chemical structures into method calls. Therefore software developers do not require extensive knowledge about chemistry and the underlying database cartridge. This decreases application development time. Molecule Database Framework is written in Java and I created it by integrating existing free and open-source tools and frameworks. The core functionality includes:•Support for multi-component compounds (mixtures)•Import and export of SD-files•Optional security (authorization)For chemical structure searching Molecule Database Framework leverages the capabilities of the Bingo Cartridge for PostgreSQL and provides type-safe searching, caching, transactions and optional method level security. Molecule Database Framework supports multi-component chemical compounds (mixtures).Furthermore the design of entity classes and the reasoning behind it are explained. By means of a simple web application I describe how the framework could be used. I then benchmarked this example application to create some basic performance expectations for chemical structure searches and import and export of SD-files. By using a simple web application it was shown that Molecule Database Framework

  19. A multi-agent conversational system with heterogeneous data sources access

    KAUST Repository

    Eisman, Eduardo M.; Navarro, Marí a; Castro, Juan Luis

    2016-01-01

    In many of the problems that can be found nowadays, information is scattered across different heterogeneous data sources. Most of the natural language interfaces just focus on a very specific part of the problem (e.g. an interface to a relational database, or an interface to an ontology). However, from the point of view of users, it does not matter where the information is stored, they just want to get the knowledge in an integrated, transparent, efficient, effective, and pleasant way. To solve this problem, this article proposes a generic multi-agent conversational architecture that follows the divide and conquer philosophy and considers two different types of agents. Expert agents are specialized in accessing different knowledge sources, and decision agents coordinate them to provide a coherent final answer to the user. This architecture has been used to design and implement SmartSeller, a specific system which includes a Virtual Assistant to answer general questions and a Bookseller to query a book database. A deep analysis regarding other relevant systems has demonstrated that our proposal provides several improvements at some key features presented along the paper.

  20. A multi-agent conversational system with heterogeneous data sources access

    KAUST Repository

    Eisman, Eduardo M.

    2016-01-28

    In many of the problems that can be found nowadays, information is scattered across different heterogeneous data sources. Most of the natural language interfaces just focus on a very specific part of the problem (e.g. an interface to a relational database, or an interface to an ontology). However, from the point of view of users, it does not matter where the information is stored, they just want to get the knowledge in an integrated, transparent, efficient, effective, and pleasant way. To solve this problem, this article proposes a generic multi-agent conversational architecture that follows the divide and conquer philosophy and considers two different types of agents. Expert agents are specialized in accessing different knowledge sources, and decision agents coordinate them to provide a coherent final answer to the user. This architecture has been used to design and implement SmartSeller, a specific system which includes a Virtual Assistant to answer general questions and a Bookseller to query a book database. A deep analysis regarding other relevant systems has demonstrated that our proposal provides several improvements at some key features presented along the paper.