WorldWideScience

Sample records for fluxes gsstf dataset

  1. Surface Turbulent Fluxes, 1x1 deg Daily Grid, Set1 V2c (GSSTF) at GES DISC

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c (GSSTF2c) Dataset recently produced through a MEaSUREs funded project led by Dr....

  2. Goddard Satellite-Based Surface Turbulent Fluxes, 0.25x0.25 deg, Daily Grid, V3, (GSSTF_F14) V3

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are part of the Goddard Satellite-based Surface Turbulent Fluxes Version 3 (GSSTF3) Dataset recently produced through a MEaSURES funded project led by Dr....

  3. Surface Turbulent Fluxes, 1x1 deg Daily Grid, Set1 V2c

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c (GSSTF2c) Dataset recently produced through a MEaSUREs funded project led by Dr....

  4. Surface Turbulent Fluxes, 1x1 deg Daily Grid, Satellite F15 V2c

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are part of the Goddard Satellite-based Surface Turbulent Fluxes Version-2c (GSSTF 2c) Dataset recently produced through a MEaSURES funded project led by...

  5. Synthesizing Global and Local Datasets to Estimate Jurisdictional Forest Carbon Fluxes in Berau, Indonesia.

    Science.gov (United States)

    Griscom, Bronson W; Ellis, Peter W; Baccini, Alessandro; Marthinus, Delon; Evans, Jeffrey S; Ruslandi

    2016-01-01

    Forest conservation efforts are increasingly being implemented at the scale of sub-national jurisdictions in order to mitigate global climate change and provide other ecosystem services. We see an urgent need for robust estimates of historic forest carbon emissions at this scale, as the basis for credible measures of climate and other benefits achieved. Despite the arrival of a new generation of global datasets on forest area change and biomass, confusion remains about how to produce credible jurisdictional estimates of forest emissions. We demonstrate a method for estimating the relevant historic forest carbon fluxes within the Regency of Berau in eastern Borneo, Indonesia. Our method integrates best available global and local datasets, and includes a comprehensive analysis of uncertainty at the regency scale. We find that Berau generated 8.91 ± 1.99 million tonnes of net CO2 emissions per year during 2000-2010. Berau is an early frontier landscape where gross emissions are 12 times higher than gross sequestration. Yet most (85%) of Berau's original forests are still standing. The majority of net emissions were due to conversion of native forests to unspecified agriculture (43% of total), oil palm (28%), and fiber plantations (9%). Most of the remainder was due to legal commercial selective logging (17%). Our overall uncertainty estimate offers an independent basis for assessing three other estimates for Berau. Two other estimates were above the upper end of our uncertainty range. We emphasize the importance of including an uncertainty range for all parameters of the emissions equation to generate a comprehensive uncertainty estimate-which has not been done before. We believe comprehensive estimates of carbon flux uncertainty are increasingly important as national and international institutions are challenged with comparing alternative estimates and identifying a credible range of historic emissions values.

  6. Towards closure of regional heat budgets in the North Atlantic using Argo floats and surface flux datasets

    Directory of Open Access Journals (Sweden)

    N. C. Wells

    2009-04-01

    Full Text Available The upper ocean heat budget (0–300 m of the North Atlantic from 20°–60° N is investigated using data from Argo profiling floats for 1999–2005 and the NCEP/NCAR and NOC surface flux datasets. Estimates of the different terms in the budget (heat storage, advection, diffusion and surface exchange are obtained using the methodology developed by Hadfield et al. (2007a, b. The method includes optimal interpolation of the individual profiles to produce gridded fields with error estimates at a 10°×10° grid box resolution. Closure of the heat budget is obtained within the error estimates for some regions – particularly the eastern subtropical Atlantic – but not for those boxes that include the Gulf Stream. Over the whole range considered, closure is obtained for 13 (9 out of 20 boxes with the NOC (NCEP/NCAR surface fluxes. The seasonal heat budget at 20–30° N, 35–25° W is considered in detail. Here, the NCEP based budget has an annual mean residual of −55±35 Wm−2 compared with a NOC based value of −4±35 Wm−2. For this box, the net heat divergence of 36 Wm−2 (Ekman=−4 Wm−2, geostrophic=11 Wm−2, diffusion=29 Wm−2 offsets the net heating of 32 Wm−2 from the NOC surface heat fluxes. The results in this box are consistent with an earlier evaluation of the fluxes using measurements from research buoys in the subduction array which revealed biases in NCEP but good agreement of the buoy values with the NOC fields.

  7. Expanding dryland ecosystem flux datasets enable novel quantification of water availability and carbon exchange in Southwestern North America

    Science.gov (United States)

    Biederman, J. A.; Scott, R. L.; Smith, W. K.; Litvak, M. E.; MacBean, N.

    2017-12-01

    Global-scale studies suggest that water-limited dryland ecosystems dominate the increasing trend in magnitude and interannual variability of the land CO2 sink. However, the terrestrial biosphere models and remote sensing models used in large-scale analyses are poorly constrained by flux measurements in drylands, which are under-represented in global datasets. In this talk, I will address this gap with eddy covariance data from 30 ecosystems across the Southwest of North America with observed ranges in annual precipitation of 100 - 1000 mm, annual temperatures of 2 - 25 °C, and records of 3 - 10 years each (160 site-years). This extensive dryland dataset enables new approaches including 1) separation of temporal and spatial patterns to infer fast and slow ecosystem responses to change, and 2) partitioning of precipitation into hydrologic losses, evaporation, and ecosystem-available water. I will then compare direct flux measurements with models and remote sensing used to scale fluxes regionally. Combining eddy covariance and streamflow measurements, I will show how evapotranspiration (ET), which is the efflux of soil moisture remaining after hydrologic losses, is a better metric than precipitation of water available to drive ecosystem CO2 exchange. Furthermore, I will present a novel method to partition ET into evaporation and transpiration using the tight coupling of transpiration and photosynthesis. In contrast with typical carbon sink function in wetter, more-studied regions, dryland sites express an annual net carbon uptake varying from -350 to +330 gC m-2. Due to less respiration losses relative to photosynthesis gains during winter, declines in winter precipitation across the Southwest since 1999 are reducing annual net CO2 uptake. Interannual variability of net uptake is larger than for wetter regions, and half the sites pivot between sinks in wet years to sources in dry years. Biospheric and remote sensing models capture only 20-30 % of interannual

  8. Flux

    DEFF Research Database (Denmark)

    Ravn, Ib

    . FLUX betegner en flyden eller strømmen, dvs. dynamik. Forstår man livet som proces og udvikling i stedet for som ting og mekanik, får man et andet billede af det gode liv end det, som den velkendte vestlige mekanicisme lægger op til. Dynamisk forstået indebærer det gode liv den bedst mulige...... kanalisering af den flux eller energi, der strømmer igennem os og giver sig til kende i vore daglige aktiviteter. Skal vores tanker, handlinger, arbejde, samvær og politiske liv organiseres efter stramme og faste regelsæt, uden slinger i valsen? Eller skal de tværtimod forløbe ganske uhindret af regler og bånd...

  9. Proteomics dataset

    DEFF Research Database (Denmark)

    Bennike, Tue Bjerg; Carlsen, Thomas Gelsing; Ellingsen, Torkell

    2017-01-01

    The datasets presented in this article are related to the research articles entitled “Neutrophil Extracellular Traps in Ulcerative Colitis: A Proteome Analysis of Intestinal Biopsies” (Bennike et al., 2015 [1]), and “Proteome Analysis of Rheumatoid Arthritis Gut Mucosa” (Bennike et al., 2017 [2])...... been deposited to the ProteomeXchange Consortium via the PRIDE partner repository with the dataset identifiers PXD001608 for ulcerative colitis and control samples, and PXD003082 for rheumatoid arthritis samples....

  10. Unified Scaling Law for flux pinning in practical superconductors: III. Minimum datasets, core parameters, and application of the Extrapolative Scaling Expression

    Science.gov (United States)

    Ekin, Jack W.; Cheggour, Najib; Goodrich, Loren; Splett, Jolene

    2017-03-01

    In Part 2 of these articles, an extensive analysis of pinning-force curves and raw scaling data was used to derive the Extrapolative Scaling Expression (ESE). This is a parameterization of the Unified Scaling Law (USL) that has the extrapolation capability of fundamental unified scaling, coupled with the application ease of a simple fitting equation. Here in Part 3, the accuracy of the ESE relation to interpolate and extrapolate limited critical-current data to obtain complete I c(B,T,ɛ) datasets is evaluated and compared with present fitting equations. Accuracy is analyzed in terms of root mean square (RMS) error and fractional deviation statistics. Highlights from 92 test cases are condensed and summarized, covering most fitting protocols and proposed parameterizations of the USL. The results show that ESE reliably extrapolates critical currents at fields B, temperatures T, and strains ɛ that are remarkably different from the fitted minimum dataset. Depending on whether the conductor is moderate-J c or high-J c, effective RMS extrapolation errors for ESE are in the range 2-5 A at 12 T, which approaches the I c measurement error (1-2%). The minimum dataset for extrapolating full I c(B,T,ɛ) characteristics is also determined from raw scaling data. It consists of one set of I c(B,ɛ) data at a fixed temperature (e.g., liquid helium temperature), and one set of I c(B,T) data at a fixed strain (e.g., zero applied strain). Error analysis of extrapolations from the minimum dataset with different fitting equations shows that ESE reduces the percentage extrapolation errors at individual data points at high fields, temperatures, and compressive strains down to 1/10th to 1/40th the size of those for extrapolations with present fitting equations. Depending on the conductor, percentage fitting errors for interpolations are also reduced to as little as 1/15th the size. The extrapolation accuracy of the ESE relation offers the prospect of straightforward implementation of

  11. Proteomics dataset

    DEFF Research Database (Denmark)

    Bennike, Tue Bjerg; Carlsen, Thomas Gelsing; Ellingsen, Torkell

    2017-01-01

    patients (Morgan et al., 2012; Abraham and Medzhitov, 2011; Bennike, 2014) [8–10. Therefore, we characterized the proteome of colon mucosa biopsies from 10 inflammatory bowel disease ulcerative colitis (UC) patients, 11 gastrointestinal healthy rheumatoid arthritis (RA) patients, and 10 controls. We...... been deposited to the ProteomeXchange Consortium via the PRIDE partner repository with the dataset identifiers PXD001608 for ulcerative colitis and control samples, and PXD003082 for rheumatoid arthritis samples....

  12. An inter-comparison of six latent and sensible heat flux products over the Southern Ocean

    Directory of Open Access Journals (Sweden)

    Lejiang Yu

    2011-11-01

    Full Text Available The latent heat fluxes (LHF and sensible heat fluxes (SHF over the Southern Ocean from six different data sets are inter-compared for the period 1988–2000. The six data sets include three satellite-based products, namely, the second version of the Goddard Satellite-Based Surface Turbulent Fluxes data set (GSSTF-2, the third version of the Hamburg Ocean Atmosphere Parameters and Fluxes from Satellite Data (HOAPS-3 and the Japanese Ocean Fluxes Data Sets with Use of Remote Sensing Observations (J-OFURO; two global reanalysis products, namely, the National Centers for Environmental Prediction–Department of Energy Reanalysis 2 data set (NCEP-2 and the European Centre for Medium-Range Weather Forecasts 40 Year Re-analysis data set (ERA-40; and the Objectively Analyzed Air–Sea Fluxes for the Global Oceans data set (OAFlux. All these products reveal a similar pattern in the averaged flux fields. The zonal mean LHF fields all exhibit a continuous increase equatorward. With an exception of HOAPS-3, the zonal mean SHF fields display a minimum value near 50°S, increasing both pole- and equatorward. The differences in the standard deviation for LHF are larger among the six data products than the differences for SHF. Over the regions where the surface fluxes are significantly influenced by the Antarctic Oscillation and the Pacific–South American teleconnection, the values and distributions of both LHF and SHF are consistent among the six products. It was found that the spatial patterns of the standard deviations and trends of LHF and SHF can be explained primarily by sea–air specific humidity and temperature differences; wind speed plays a minor role.

  13. EPA Nanorelease Dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — EPA Nanorelease Dataset. This dataset is associated with the following publication: Wohlleben, W., C. Kingston, J. Carter, E. Sahle-Demessie, S. Vazquez-Campos, B....

  14. Aaron Journal article datasets

    Data.gov (United States)

    U.S. Environmental Protection Agency — All figures used in the journal article are in netCDF format. This dataset is associated with the following publication: Sims, A., K. Alapaty , and S. Raman....

  15. Integrated Surface Dataset (Global)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The Integrated Surface (ISD) Dataset (ISD) is composed of worldwide surface weather observations from over 35,000 stations, though the best spatial coverage is...

  16. Control Measure Dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — The EPA Control Measure Dataset is a collection of documents describing air pollution control available to regulated facilities for the control and abatement of air...

  17. National Hydrography Dataset (NHD)

    Data.gov (United States)

    Kansas Data Access and Support Center — The National Hydrography Dataset (NHD) is a feature-based database that interconnects and uniquely identifies the stream segments or reaches that comprise the...

  18. Market Squid Ecology Dataset

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This dataset contains ecological information collected on the major adult spawning and juvenile habitats of market squid off California and the US Pacific Northwest....

  19. Tables and figure datasets

    Data.gov (United States)

    U.S. Environmental Protection Agency — Soil and air concentrations of asbestos in Sumas study. This dataset is associated with the following publication: Wroble, J., T. Frederick, A. Frame, and D....

  20. Isfahan MISP Dataset.

    Science.gov (United States)

    Kashefpur, Masoud; Kafieh, Rahele; Jorjandi, Sahar; Golmohammadi, Hadis; Khodabande, Zahra; Abbasi, Mohammadreza; Teifuri, Nilufar; Fakharzadeh, Ali Akbar; Kashefpoor, Maryam; Rabbani, Hossein

    2017-01-01

    An online depository was introduced to share clinical ground truth with the public and provide open access for researchers to evaluate their computer-aided algorithms. PHP was used for web programming and MySQL for database managing. The website was entitled "biosigdata.com." It was a fast, secure, and easy-to-use online database for medical signals and images. Freely registered users could download the datasets and could also share their own supplementary materials while maintaining their privacies (citation and fee). Commenting was also available for all datasets, and automatic sitemap and semi-automatic SEO indexing have been set for the site. A comprehensive list of available websites for medical datasets is also presented as a Supplementary (http://journalonweb.com/tempaccess/4800.584.JMSS_55_16I3253.pdf).

  1. Mridangam stroke dataset

    OpenAIRE

    CompMusic

    2014-01-01

    The audio examples were recorded from a professional Carnatic percussionist in a semi-anechoic studio conditions by Akshay Anantapadmanabhan using SM-58 microphones and an H4n ZOOM recorder. The audio was sampled at 44.1 kHz and stored as 16 bit wav files. The dataset can be used for training models for each Mridangam stroke. /n/nA detailed description of the Mridangam and its strokes can be found in the paper below. A part of the dataset was used in the following paper. /nAkshay Anantapadman...

  2. The GTZAN dataset

    DEFF Research Database (Denmark)

    Sturm, Bob L.

    2013-01-01

    The GTZAN dataset appears in at least 100 published works, and is the most-used public dataset for evaluation in machine listening research for music genre recognition (MGR). Our recent work, however, shows GTZAN has several faults (repetitions, mislabelings, and distortions), which challenge...... of GTZAN, and provide a catalog of its faults. We review how GTZAN has been used in MGR research, and find few indications that its faults have been known and considered. Finally, we rigorously study the effects of its faults on evaluating five different MGR systems. The lesson is not to banish GTZAN...

  3. Dataset - Adviesregel PPL 2010

    NARCIS (Netherlands)

    Evert, van F.K.; Schans, van der D.A.; Geel, van W.C.A.; Slabbekoorn, J.J.; Booij, R.; Jukema, J.N.; Meurs, E.J.J.; Uenk, D.

    2011-01-01

    This dataset contains experimental data from a number of field experiments with potato in The Netherlands (Van Evert et al., 2011). The data are presented as an SQL dump of a PostgreSQL database (version 8.4.4). An outline of the entity-relationship diagram of the database is given in an

  4. National Elevation Dataset

    Science.gov (United States)

    ,

    2002-01-01

    The National Elevation Dataset (NED) is a new raster product assembled by the U.S. Geological Survey. NED is designed to provide National elevation data in a seamless form with a consistent datum, elevation unit, and projection. Data corrections were made in the NED assembly process to minimize artifacts, perform edge matching, and fill sliver areas of missing data. NED has a resolution of one arc-second (approximately 30 meters) for the conterminous United States, Hawaii, Puerto Rico and the island territories and a resolution of two arc-seconds for Alaska. NED data sources have a variety of elevation units, horizontal datums, and map projections. In the NED assembly process the elevation values are converted to decimal meters as a consistent unit of measure, NAD83 is consistently used as horizontal datum, and all the data are recast in a geographic projection. Older DEM's produced by methods that are now obsolete have been filtered during the NED assembly process to minimize artifacts that are commonly found in data produced by these methods. Artifact removal greatly improves the quality of the slope, shaded-relief, and synthetic drainage information that can be derived from the elevation data. Figure 2 illustrates the results of this artifact removal filtering. NED processing also includes steps to adjust values where adjacent DEM's do not match well, and to fill sliver areas of missing data between DEM's. These processing steps ensure that NED has no void areas and artificial discontinuities have been minimized. The artifact removal filtering process does not eliminate all of the artifacts. In areas where the only available DEM is produced by older methods, then "striping" may still occur.

  5. NP-PAH Interaction Dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — Dataset presents concentrations of organic pollutants, such as polyaromatic hydrocarbon compounds, in water samples. Water samples of known volume and concentration...

  6. Editorial: Datasets for Learning Analytics

    NARCIS (Netherlands)

    Dietze, Stefan; George, Siemens; Davide, Taibi; Drachsler, Hendrik

    2018-01-01

    The European LinkedUp and LACE (Learning Analytics Community Exchange) project have been responsible for setting up a series of data challenges at the LAK conferences 2013 and 2014 around the LAK dataset. The LAK datasets consists of a rich collection of full text publications in the domain of

  7. Open University Learning Analytics dataset.

    Science.gov (United States)

    Kuzilek, Jakub; Hlosta, Martin; Zdrahal, Zdenek

    2017-11-28

    Learning Analytics focuses on the collection and analysis of learners' data to improve their learning experience by providing informed guidance and to optimise learning materials. To support the research in this area we have developed a dataset, containing data from courses presented at the Open University (OU). What makes the dataset unique is the fact that it contains demographic data together with aggregated clickstream data of students' interactions in the Virtual Learning Environment (VLE). This enables the analysis of student behaviour, represented by their actions. The dataset contains the information about 22 courses, 32,593 students, their assessment results, and logs of their interactions with the VLE represented by daily summaries of student clicks (10,655,280 entries). The dataset is freely available at https://analyse.kmi.open.ac.uk/open_dataset under a CC-BY 4.0 license.

  8. Turkey Run Landfill Emissions Dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — landfill emissions measurements for the Turkey run landfill in Georgia. This dataset is associated with the following publication: De la Cruz, F., R. Green, G....

  9. Dataset of NRDA emission data

    Data.gov (United States)

    U.S. Environmental Protection Agency — Emissions data from open air oil burns. This dataset is associated with the following publication: Gullett, B., J. Aurell, A. Holder, B. Mitchell, D. Greenwell, M....

  10. Chemical product and function dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — Merged product weight fraction and chemical function data. This dataset is associated with the following publication: Isaacs , K., M. Goldsmith, P. Egeghy , K....

  11. The NOAA Dataset Identifier Project

    Science.gov (United States)

    de la Beaujardiere, J.; Mccullough, H.; Casey, K. S.

    2013-12-01

    The US National Oceanic and Atmospheric Administration (NOAA) initiated a project in 2013 to assign persistent identifiers to datasets archived at NOAA and to create informational landing pages about those datasets. The goals of this project are to enable the citation of datasets used in products and results in order to help provide credit to data producers, to support traceability and reproducibility, and to enable tracking of data usage and impact. A secondary goal is to encourage the submission of datasets for long-term preservation, because only archived datasets will be eligible for a NOAA-issued identifier. A team was formed with representatives from the National Geophysical, Oceanographic, and Climatic Data Centers (NGDC, NODC, NCDC) to resolve questions including which identifier scheme to use (answer: Digital Object Identifier - DOI), whether or not to embed semantics in identifiers (no), the level of granularity at which to assign identifiers (as coarsely as reasonable), how to handle ongoing time-series data (do not break into chunks), creation mechanism for the landing page (stylesheet from formal metadata record preferred), and others. Decisions made and implementation experience gained will inform the writing of a Data Citation Procedural Directive to be issued by the Environmental Data Management Committee in 2014. Several identifiers have been issued as of July 2013, with more on the way. NOAA is now reporting the number as a metric to federal Open Government initiatives. This paper will provide further details and status of the project.

  12. The Harvard organic photovoltaic dataset.

    Science.gov (United States)

    Lopez, Steven A; Pyzer-Knapp, Edward O; Simm, Gregor N; Lutzow, Trevor; Li, Kewei; Seress, Laszlo R; Hachmann, Johannes; Aspuru-Guzik, Alán

    2016-09-27

    The Harvard Organic Photovoltaic Dataset (HOPV15) presented in this work is a collation of experimental photovoltaic data from the literature, and corresponding quantum-chemical calculations performed over a range of conformers, each with quantum chemical results using a variety of density functionals and basis sets. It is anticipated that this dataset will be of use in both relating electronic structure calculations to experimental observations through the generation of calibration schemes, as well as for the creation of new semi-empirical methods and the benchmarking of current and future model chemistries for organic electronic applications.

  13. The Harvard organic photovoltaic dataset

    Science.gov (United States)

    Lopez, Steven A.; Pyzer-Knapp, Edward O.; Simm, Gregor N.; Lutzow, Trevor; Li, Kewei; Seress, Laszlo R.; Hachmann, Johannes; Aspuru-Guzik, Alán

    2016-01-01

    The Harvard Organic Photovoltaic Dataset (HOPV15) presented in this work is a collation of experimental photovoltaic data from the literature, and corresponding quantum-chemical calculations performed over a range of conformers, each with quantum chemical results using a variety of density functionals and basis sets. It is anticipated that this dataset will be of use in both relating electronic structure calculations to experimental observations through the generation of calibration schemes, as well as for the creation of new semi-empirical methods and the benchmarking of current and future model chemistries for organic electronic applications. PMID:27676312

  14. Querying Large Biological Network Datasets

    Science.gov (United States)

    Gulsoy, Gunhan

    2013-01-01

    New experimental methods has resulted in increasing amount of genetic interaction data to be generated every day. Biological networks are used to store genetic interaction data gathered. Increasing amount of data available requires fast large scale analysis methods. Therefore, we address the problem of querying large biological network datasets.…

  15. Fluxnet Synthesis Dataset Collaboration Infrastructure

    Energy Technology Data Exchange (ETDEWEB)

    Agarwal, Deborah A. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Humphrey, Marty [Univ. of Virginia, Charlottesville, VA (United States); van Ingen, Catharine [Microsoft. San Francisco, CA (United States); Beekwilder, Norm [Univ. of Virginia, Charlottesville, VA (United States); Goode, Monte [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Jackson, Keith [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Rodriguez, Matt [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Weber, Robin [Univ. of California, Berkeley, CA (United States)

    2008-02-06

    The Fluxnet synthesis dataset originally compiled for the La Thuile workshop contained approximately 600 site years. Since the workshop, several additional site years have been added and the dataset now contains over 920 site years from over 240 sites. A data refresh update is expected to increase those numbers in the next few months. The ancillary data describing the sites continues to evolve as well. There are on the order of 120 site contacts and 60proposals have been approved to use thedata. These proposals involve around 120 researchers. The size and complexity of the dataset and collaboration has led to a new approach to providing access to the data and collaboration support and the support team attended the workshop and worked closely with the attendees and the Fluxnet project office to define the requirements for the support infrastructure. As a result of this effort, a new website (http://www.fluxdata.org) has been created to provide access to the Fluxnet synthesis dataset. This new web site is based on a scientific data server which enables browsing of the data on-line, data download, and version tracking. We leverage database and data analysis tools such as OLAP data cubes and web reports to enable browser and Excel pivot table access to the data.

  16. CERC Dataset (Full Hadza Data)

    DEFF Research Database (Denmark)

    2016-01-01

    The dataset includes demographic, behavioral, and religiosity data from eight different populations from around the world. The samples were drawn from: (1) Coastal and (2) Inland Tanna, Vanuatu; (3) Hadzaland, Tanzania; (4) Lovu, Fiji; (5) Pointe aux Piment, Mauritius; (6) Pesqueiro, Brazil; (7......) Kyzyl, Tyva Republic; and (8) Yasawa, Fiji. Related publication: Purzycki, et al. (2016). Moralistic Gods, Supernatural Punishment and the Expansion of Human Sociality. Nature, 530(7590): 327-330....

  17. Viking Seismometer PDS Archive Dataset

    Science.gov (United States)

    Lorenz, R. D.

    2016-12-01

    The Viking Lander 2 seismometer operated successfully for over 500 Sols on the Martian surface, recording at least one likely candidate Marsquake. The Viking mission, in an era when data handling hardware (both on board and on the ground) was limited in capability, predated modern planetary data archiving, and ad-hoc repositories of the data, and the very low-level record at NSSDC, were neither convenient to process nor well-known. In an effort supported by the NASA Mars Data Analysis Program, we have converted the bulk of the Viking dataset (namely the 49,000 and 270,000 records made in High- and Event- modes at 20 and 1 Hz respectively) into a simple ASCII table format. Additionally, since wind-generated lander motion is a major component of the signal, contemporaneous meteorological data are included in summary records to facilitate correlation. These datasets are being archived at the PDS Geosciences Node. In addition to brief instrument and dataset descriptions, the archive includes code snippets in the freely-available language 'R' to demonstrate plotting and analysis. Further, we present examples of lander-generated noise, associated with the sampler arm, instrument dumps and other mechanical operations.

  18. PHYSICS PERFORMANCE AND DATASET (PPD)

    CERN Multimedia

    L. Silvestris

    2013-01-01

    The first part of the Long Shutdown period has been dedicated to the preparation of the samples for the analysis targeting the summer conferences. In particular, the 8 TeV data acquired in 2012, including most of the “parked datasets”, have been reconstructed profiting from improved alignment and calibration conditions for all the sub-detectors. A careful planning of the resources was essential in order to deliver the datasets well in time to the analysts, and to schedule the update of all the conditions and calibrations needed at the analysis level. The newly reprocessed data have undergone detailed scrutiny by the Dataset Certification team allowing to recover some of the data for analysis usage and further improving the certification efficiency, which is now at 91% of the recorded luminosity. With the aim of delivering a consistent dataset for 2011 and 2012, both in terms of conditions and release (53X), the PPD team is now working to set up a data re-reconstruction and a new MC pro...

  19. RARD: The Related-Article Recommendation Dataset

    OpenAIRE

    Beel, Joeran; Carevic, Zeljko; Schaible, Johann; Neusch, Gabor

    2017-01-01

    Recommender-system datasets are used for recommender-system evaluations, training machine-learning algorithms, and exploring user behavior. While there are many datasets for recommender systems in the domains of movies, books, and music, there are rather few datasets from research-paper recommender systems. In this paper, we introduce RARD, the Related-Article Recommendation Dataset, from the digital library Sowiport and the recommendation-as-a-service provider Mr. DLib. The dataset contains ...

  20. Passive Containment DataSet

    Science.gov (United States)

    This data is for Figures 6 and 7 in the journal article. The data also includes the two EPANET input files used for the analysis described in the paper, one for the looped system and one for the block system.This dataset is associated with the following publication:Grayman, W., R. Murray , and D. Savic. Redesign of Water Distribution Systems for Passive Containment of Contamination. JOURNAL OF THE AMERICAN WATER WORKS ASSOCIATION. American Water Works Association, Denver, CO, USA, 108(7): 381-391, (2016).

  1. The CMS dataset bookkeeping service

    Science.gov (United States)

    Afaq, A.; Dolgert, A.; Guo, Y.; Jones, C.; Kosyakov, S.; Kuznetsov, V.; Lueking, L.; Riley, D.; Sekhri, V.

    2008-07-01

    The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event data from Monte Carlo and Detector sources. It provides the ability to identify MC or trigger source, track data provenance, construct datasets for analysis, and discover interesting data. CMS requires processing and analysis activities at various service levels and the DBS system provides support for localized processing or private analysis, as well as global access for CMS users at large. Catalog entries can be moved among the various service levels with a simple set of migration tools, thus forming a loose federation of databases. DBS is available to CMS users via a Python API, Command Line, and a Discovery web page interfaces. The system is built as a multi-tier web application with Java servlets running under Tomcat, with connections via JDBC to Oracle or MySQL database backends. Clients connect to the service through HTTP or HTTPS with authentication provided by GRID certificates and authorization through VOMS. DBS is an integral part of the overall CMS Data Management and Workflow Management systems.

  2. The CMS dataset bookkeeping service

    Energy Technology Data Exchange (ETDEWEB)

    Afaq, A; Guo, Y; Kosyakov, S; Lueking, L; Sekhri, V [Fermilab, Batavia, Illinois 60510 (United States); Dolgert, A; Jones, C; Kuznetsov, V; Riley, D [Cornell University, Ithaca, New York 14850 (United States)

    2008-07-15

    The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event data from Monte Carlo and Detector sources. It provides the ability to identify MC or trigger source, track data provenance, construct datasets for analysis, and discover interesting data. CMS requires processing and analysis activities at various service levels and the DBS system provides support for localized processing or private analysis, as well as global access for CMS users at large. Catalog entries can be moved among the various service levels with a simple set of migration tools, thus forming a loose federation of databases. DBS is available to CMS users via a Python API, Command Line, and a Discovery web page interfaces. The system is built as a multi-tier web application with Java servlets running under Tomcat, with connections via JDBC to Oracle or MySQL database backends. Clients connect to the service through HTTP or HTTPS with authentication provided by GRID certificates and authorization through VOMS. DBS is an integral part of the overall CMS Data Management and Workflow Management systems.

  3. The CMS dataset bookkeeping service

    International Nuclear Information System (INIS)

    Afaq, A; Guo, Y; Kosyakov, S; Lueking, L; Sekhri, V; Dolgert, A; Jones, C; Kuznetsov, V; Riley, D

    2008-01-01

    The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event data from Monte Carlo and Detector sources. It provides the ability to identify MC or trigger source, track data provenance, construct datasets for analysis, and discover interesting data. CMS requires processing and analysis activities at various service levels and the DBS system provides support for localized processing or private analysis, as well as global access for CMS users at large. Catalog entries can be moved among the various service levels with a simple set of migration tools, thus forming a loose federation of databases. DBS is available to CMS users via a Python API, Command Line, and a Discovery web page interfaces. The system is built as a multi-tier web application with Java servlets running under Tomcat, with connections via JDBC to Oracle or MySQL database backends. Clients connect to the service through HTTP or HTTPS with authentication provided by GRID certificates and authorization through VOMS. DBS is an integral part of the overall CMS Data Management and Workflow Management systems

  4. The CMS dataset bookkeeping service

    International Nuclear Information System (INIS)

    Afaq, Anzar; Dolgert, Andrew; Guo, Yuyi; Jones, Chris; Kosyakov, Sergey; Kuznetsov, Valentin; Lueking, Lee; Riley, Dan; Sekhri, Vijay

    2007-01-01

    The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event data from Monte Carlo and Detector sources. It provides the ability to identify MC or trigger source, track data provenance, construct datasets for analysis, and discover interesting data. CMS requires processing and analysis activities at various service levels and the DBS system provides support for localized processing or private analysis, as well as global access for CMS users at large. Catalog entries can be moved among the various service levels with a simple set of migration tools, thus forming a loose federation of databases. DBS is available to CMS users via a Python API, Command Line, and a Discovery web page interfaces. The system is built as a multi-tier web application with Java servlets running under Tomcat, with connections via JDBC to Oracle or MySQL database backends. Clients connect to the service through HTTP or HTTPS with authentication provided by GRID certificates and authorization through VOMS. DBS is an integral part of the overall CMS Data Management and Workflow Management systems

  5. 2008 TIGER/Line Nationwide Dataset

    Data.gov (United States)

    California Natural Resource Agency — This dataset contains a nationwide build of the 2008 TIGER/Line datasets from the US Census Bureau downloaded in April 2009. The TIGER/Line Shapefiles are an extract...

  6. Satellite-Based Precipitation Datasets

    Science.gov (United States)

    Munchak, S. J.; Huffman, G. J.

    2017-12-01

    Of the possible sources of precipitation data, those based on satellites provide the greatest spatial coverage. There is a wide selection of datasets, algorithms, and versions from which to choose, which can be confusing to non-specialists wishing to use the data. The International Precipitation Working Group (IPWG) maintains tables of the major publicly available, long-term, quasi-global precipitation data sets (http://www.isac.cnr.it/ ipwg/data/datasets.html), and this talk briefly reviews the various categories. As examples, NASA provides two sets of quasi-global precipitation data sets: the older Tropical Rainfall Measuring Mission (TRMM) Multi-satellite Precipitation Analysis (TMPA) and current Integrated Multi-satellitE Retrievals for Global Precipitation Measurement (GPM) mission (IMERG). Both provide near-real-time and post-real-time products that are uniformly gridded in space and time. The TMPA products are 3-hourly 0.25°x0.25° on the latitude band 50°N-S for about 16 years, while the IMERG products are half-hourly 0.1°x0.1° on 60°N-S for over 3 years (with plans to go to 16+ years in Spring 2018). In addition to the precipitation estimates, each data set provides fields of other variables, such as the satellite sensor providing estimates and estimated random error. The discussion concludes with advice about determining suitability for use, the necessity of being clear about product names and versions, and the need for continued support for satellite- and surface-based observation.

  7. Critical flux determination by flux-stepping

    DEFF Research Database (Denmark)

    Beier, Søren; Jonsson, Gunnar Eigil

    2010-01-01

    In membrane filtration related scientific literature, often step-by-step determined critical fluxes are reported. Using a dynamic microfiltration device, it is shown that critical fluxes determined from two different flux-stepping methods are dependent upon operational parameters such as step...... length, step height, and.flux start level. Filtrating 8 kg/m(3) yeast cell suspensions by a vibrating 0.45 x 10(-6) m pore size microfiltration hollow fiber module, critical fluxes from 5.6 x 10(-6) to 1.2 x 10(-5) m/s have been measured using various step lengths from 300 to 1200 seconds. Thus......, such values are more or less useless in itself as critical flux predictors, and constant flux verification experiments have to be conducted to check if the determined critical fluxes call predict sustainable flux regimes. However, it is shown that using the step-by-step predicted critical fluxes as start...

  8. PHYSICS PERFORMANCE AND DATASET (PPD)

    CERN Multimedia

    L. Silvestris

    2012-01-01

      Introduction The first part of the year presented an important test for the new Physics Performance and Dataset (PPD) group (cf. its mandate: http://cern.ch/go/8f77). The activity was focused on the validation of the new releases meant for the Monte Carlo (MC) production and the data-processing in 2012 (CMSSW 50X and 52X), and on the preparation of the 2012 operations. In view of the Chamonix meeting, the PPD and physics groups worked to understand the impact of the higher pile-up scenario on some of the flagship Higgs analyses to better quantify the impact of the high luminosity on the CMS physics potential. A task force is working on the optimisation of the reconstruction algorithms and on the code to cope with the performance requirements imposed by the higher event occupancy as foreseen for 2012. Concerning the preparation for the analysis of the new data, a new MC production has been prepared. The new samples, simulated at 8 TeV, are already being produced and the digitisation and recons...

  9. Pattern Analysis On Banking Dataset

    Directory of Open Access Journals (Sweden)

    Amritpal Singh

    2015-06-01

    Full Text Available Abstract Everyday refinement and development of technology has led to an increase in the competition between the Tech companies and their going out of way to crack the system andbreak down. Thus providing Data mining a strategically and security-wise important area for many business organizations including banking sector. It allows the analyzes of important information in the data warehouse and assists the banks to look for obscure patterns in a group and discover unknown relationship in the data.Banking systems needs to process ample amount of data on daily basis related to customer information their credit card details limit and collateral details transaction details risk profiles Anti Money Laundering related information trade finance data. Thousands of decisionsbased on the related data are taken in a bank daily. This paper analyzes the banking dataset in the weka environment for the detection of interesting patterns based on its applications ofcustomer acquisition customer retention management and marketing and management of risk fraudulence detections.

  10. PHYSICS PERFORMANCE AND DATASET (PPD)

    CERN Multimedia

    L. Silvestris

    2013-01-01

    The PPD activities, in the first part of 2013, have been focused mostly on the final physics validation and preparation for the data reprocessing of the full 8 TeV datasets with the latest calibrations. These samples will be the basis for the preliminary results for summer 2013 but most importantly for the final publications on the 8 TeV Run 1 data. The reprocessing involves also the reconstruction of a significant fraction of “parked data” that will allow CMS to perform a whole new set of precision analyses and searches. In this way the CMSSW release 53X is becoming the legacy release for the 8 TeV Run 1 data. The regular operation activities have included taking care of the prolonged proton-proton data taking and the run with proton-lead collisions that ended in February. The DQM and Data Certification team has deployed a continuous effort to promptly certify the quality of the data. The luminosity-weighted certification efficiency (requiring all sub-detectors to be certified as usab...

  11. The Geometry of Finite Equilibrium Datasets

    DEFF Research Database (Denmark)

    Balasko, Yves; Tvede, Mich

    We investigate the geometry of finite datasets defined by equilibrium prices, income distributions, and total resources. We show that the equilibrium condition imposes no restrictions if total resources are collinear, a property that is robust to small perturbations. We also show that the set...... of equilibrium datasets is pathconnected when the equilibrium condition does impose restrictions on datasets, as for example when total resources are widely non collinear....

  12. IPCC Socio-Economic Baseline Dataset

    Data.gov (United States)

    National Aeronautics and Space Administration — The Intergovernmental Panel on Climate Change (IPCC) Socio-Economic Baseline Dataset consists of population, human development, economic, water resources, land...

  13. Veterans Affairs Suicide Prevention Synthetic Dataset

    Data.gov (United States)

    Department of Veterans Affairs — The VA's Veteran Health Administration, in support of the Open Data Initiative, is providing the Veterans Affairs Suicide Prevention Synthetic Dataset (VASPSD). The...

  14. Nanoparticle-organic pollutant interaction dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — Dataset presents concentrations of organic pollutants, such as polyaromatic hydrocarbon compounds, in water samples. Water samples of known volume and concentration...

  15. An Annotated Dataset of 14 Meat Images

    DEFF Research Database (Denmark)

    Stegmann, Mikkel Bille

    2002-01-01

    This note describes a dataset consisting of 14 annotated images of meat. Points of correspondence are placed on each image. As such, the dataset can be readily used for building statistical models of shape. Further, format specifications and terms of use are given.......This note describes a dataset consisting of 14 annotated images of meat. Points of correspondence are placed on each image. As such, the dataset can be readily used for building statistical models of shape. Further, format specifications and terms of use are given....

  16. SIMADL: Simulated Activities of Daily Living Dataset

    Directory of Open Access Journals (Sweden)

    Talal Alshammari

    2018-04-01

    Full Text Available With the realisation of the Internet of Things (IoT paradigm, the analysis of the Activities of Daily Living (ADLs, in a smart home environment, is becoming an active research domain. The existence of representative datasets is a key requirement to advance the research in smart home design. Such datasets are an integral part of the visualisation of new smart home concepts as well as the validation and evaluation of emerging machine learning models. Machine learning techniques that can learn ADLs from sensor readings are used to classify, predict and detect anomalous patterns. Such techniques require data that represent relevant smart home scenarios, for training, testing and validation. However, the development of such machine learning techniques is limited by the lack of real smart home datasets, due to the excessive cost of building real smart homes. This paper provides two datasets for classification and anomaly detection. The datasets are generated using OpenSHS, (Open Smart Home Simulator, which is a simulation software for dataset generation. OpenSHS records the daily activities of a participant within a virtual environment. Seven participants simulated their ADLs for different contexts, e.g., weekdays, weekends, mornings and evenings. Eighty-four files in total were generated, representing approximately 63 days worth of activities. Forty-two files of classification of ADLs were simulated in the classification dataset and the other forty-two files are for anomaly detection problems in which anomalous patterns were simulated and injected into the anomaly detection dataset.

  17. ASSISTments Dataset from Multiple Randomized Controlled Experiments

    Science.gov (United States)

    Selent, Douglas; Patikorn, Thanaporn; Heffernan, Neil

    2016-01-01

    In this paper, we present a dataset consisting of data generated from 22 previously and currently running randomized controlled experiments inside the ASSISTments online learning platform. This dataset provides data mining opportunities for researchers to analyze ASSISTments data in a convenient format across multiple experiments at the same time.…

  18. Synthetic and Empirical Capsicum Annuum Image Dataset

    NARCIS (Netherlands)

    Barth, R.

    2016-01-01

    This dataset consists of per-pixel annotated synthetic (10500) and empirical images (50) of Capsicum annuum, also known as sweet or bell pepper, situated in a commercial greenhouse. Furthermore, the source models to generate the synthetic images are included. The aim of the datasets are to

  19. Design of an audio advertisement dataset

    Science.gov (United States)

    Fu, Yutao; Liu, Jihong; Zhang, Qi; Geng, Yuting

    2015-12-01

    Since more and more advertisements swarm into radios, it is necessary to establish an audio advertising dataset which could be used to analyze and classify the advertisement. A method of how to establish a complete audio advertising dataset is presented in this paper. The dataset is divided into four different kinds of advertisements. Each advertisement's sample is given in *.wav file format, and annotated with a txt file which contains its file name, sampling frequency, channel number, broadcasting time and its class. The classifying rationality of the advertisements in this dataset is proved by clustering the different advertisements based on Principal Component Analysis (PCA). The experimental results show that this audio advertisement dataset offers a reliable set of samples for correlative audio advertisement experimental studies.

  20. GLEAM version 3: Global Land Evaporation Datasets and Model

    Science.gov (United States)

    Martens, B.; Miralles, D. G.; Lievens, H.; van der Schalie, R.; de Jeu, R.; Fernandez-Prieto, D.; Verhoest, N.

    2015-12-01

    Terrestrial evaporation links energy, water and carbon cycles over land and is therefore a key variable of the climate system. However, the global-scale magnitude and variability of the flux, and the sensitivity of the underlying physical process to changes in environmental factors, are still poorly understood due to limitations in in situ measurements. As a result, several methods have risen to estimate global patterns of land evaporation from satellite observations. However, these algorithms generally differ in their approach to model evaporation, resulting in large differences in their estimates. One of these methods is GLEAM, the Global Land Evaporation: the Amsterdam Methodology. GLEAM estimates terrestrial evaporation based on daily satellite observations of meteorological variables, vegetation characteristics and soil moisture. Since the publication of the first version of the algorithm (2011), the model has been widely applied to analyse trends in the water cycle and land-atmospheric feedbacks during extreme hydrometeorological events. A third version of the GLEAM global datasets is foreseen by the end of 2015. Given the relevance of having a continuous and reliable record of global-scale evaporation estimates for climate and hydrological research, the establishment of an online data portal to host these data to the public is also foreseen. In this new release of the GLEAM datasets, different components of the model have been updated, with the most significant change being the revision of the data assimilation algorithm. In this presentation, we will highlight the most important changes of the methodology and present three new GLEAM datasets and their validation against in situ observations and an alternative dataset of terrestrial evaporation (ERA-Land). Results of the validation exercise indicate that the magnitude and the spatiotemporal variability of the modelled evaporation agree reasonably well with the estimates of ERA-Land and the in situ

  1. The Kinetics Human Action Video Dataset

    OpenAIRE

    Kay, Will; Carreira, Joao; Simonyan, Karen; Zhang, Brian; Hillier, Chloe; Vijayanarasimhan, Sudheendra; Viola, Fabio; Green, Tim; Back, Trevor; Natsev, Paul; Suleyman, Mustafa; Zisserman, Andrew

    2017-01-01

    We describe the DeepMind Kinetics human action video dataset. The dataset contains 400 human action classes, with at least 400 video clips for each action. Each clip lasts around 10s and is taken from a different YouTube video. The actions are human focussed and cover a broad range of classes including human-object interactions such as playing instruments, as well as human-human interactions such as shaking hands. We describe the statistics of the dataset, how it was collected, and give some ...

  2. Magnetic-flux pump

    Science.gov (United States)

    Hildebrandt, A. F.; Elleman, D. D.; Whitmore, F. C. (Inventor)

    1966-01-01

    A magnetic flux pump is described for increasing the intensity of a magnetic field by transferring flux from one location to the magnetic field. The device includes a pair of communicating cavities formed in a block of superconducting material, and a piston for displacing the trapped magnetic flux into the secondary cavity producing a field having an intense flux density.

  3. Radon flux measurement methodologies

    International Nuclear Information System (INIS)

    Nielson, K.K.; Rogers, V.C.

    1984-01-01

    Five methods for measuring radon fluxes are evaluated: the accumulator can, a small charcoal sampler, a large-area charcoal sampler, the ''Big Louie'' charcoal sampler, and the charcoal tent sampler. An experimental comparison of the five flux measurement techniques was also conducted. Excellent agreement was obtained between the measured radon fluxes and fluxes predicted from radium and emanation measurements

  4. BASE MAP DATASET, LOS ANGELES COUNTY, CALIFORNIA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  5. BASE MAP DATASET, CHEROKEE COUNTY, SOUTH CAROLINA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  6. SIAM 2007 Text Mining Competition dataset

    Data.gov (United States)

    National Aeronautics and Space Administration — Subject Area: Text Mining Description: This is the dataset used for the SIAM 2007 Text Mining competition. This competition focused on developing text mining...

  7. Harvard Aging Brain Study : Dataset and accessibility

    NARCIS (Netherlands)

    Dagley, Alexander; LaPoint, Molly; Huijbers, Willem; Hedden, Trey; McLaren, Donald G.; Chatwal, Jasmeer P.; Papp, Kathryn V.; Amariglio, Rebecca E.; Blacker, Deborah; Rentz, Dorene M.; Johnson, Keith A.; Sperling, Reisa A.; Schultz, Aaron P.

    2017-01-01

    The Harvard Aging Brain Study is sharing its data with the global research community. The longitudinal dataset consists of a 284-subject cohort with the following modalities acquired: demographics, clinical assessment, comprehensive neuropsychological testing, clinical biomarkers, and neuroimaging.

  8. BASE MAP DATASET, HONOLULU COUNTY, HAWAII, USA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  9. BASE MAP DATASET, EDGEFIELD COUNTY, SOUTH CAROLINA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  10. Simulation of Smart Home Activity Datasets

    Directory of Open Access Journals (Sweden)

    Jonathan Synnott

    2015-06-01

    Full Text Available A globally ageing population is resulting in an increased prevalence of chronic conditions which affect older adults. Such conditions require long-term care and management to maximize quality of life, placing an increasing strain on healthcare resources. Intelligent environments such as smart homes facilitate long-term monitoring of activities in the home through the use of sensor technology. Access to sensor datasets is necessary for the development of novel activity monitoring and recognition approaches. Access to such datasets is limited due to issues such as sensor cost, availability and deployment time. The use of simulated environments and sensors may address these issues and facilitate the generation of comprehensive datasets. This paper provides a review of existing approaches for the generation of simulated smart home activity datasets, including model-based approaches and interactive approaches which implement virtual sensors, environments and avatars. The paper also provides recommendation for future work in intelligent environment simulation.

  11. Simulation of Smart Home Activity Datasets.

    Science.gov (United States)

    Synnott, Jonathan; Nugent, Chris; Jeffers, Paul

    2015-06-16

    A globally ageing population is resulting in an increased prevalence of chronic conditions which affect older adults. Such conditions require long-term care and management to maximize quality of life, placing an increasing strain on healthcare resources. Intelligent environments such as smart homes facilitate long-term monitoring of activities in the home through the use of sensor technology. Access to sensor datasets is necessary for the development of novel activity monitoring and recognition approaches. Access to such datasets is limited due to issues such as sensor cost, availability and deployment time. The use of simulated environments and sensors may address these issues and facilitate the generation of comprehensive datasets. This paper provides a review of existing approaches for the generation of simulated smart home activity datasets, including model-based approaches and interactive approaches which implement virtual sensors, environments and avatars. The paper also provides recommendation for future work in intelligent environment simulation.

  12. Environmental Dataset Gateway (EDG) REST Interface

    Data.gov (United States)

    U.S. Environmental Protection Agency — Use the Environmental Dataset Gateway (EDG) to find and access EPA's environmental resources. Many options are available for easily reusing EDG content in other...

  13. BASE MAP DATASET, INYO COUNTY, OKLAHOMA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  14. BASE MAP DATASET, JACKSON COUNTY, OKLAHOMA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  15. BASE MAP DATASET, SANTA CRIZ COUNTY, CALIFORNIA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  16. Climate Prediction Center IR 4km Dataset

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — CPC IR 4km dataset was created from all available individual geostationary satellite data which have been merged to form nearly seamless global (60N-60S) IR...

  17. BASE MAP DATASET, MAYES COUNTY, OKLAHOMA, USA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications: cadastral, geodetic control,...

  18. BASE MAP DATASET, KINGFISHER COUNTY, OKLAHOMA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — FEMA Framework Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme,...

  19. Comparison of recent SnIa datasets

    International Nuclear Information System (INIS)

    Sanchez, J.C. Bueno; Perivolaropoulos, L.; Nesseris, S.

    2009-01-01

    We rank the six latest Type Ia supernova (SnIa) datasets (Constitution (C), Union (U), ESSENCE (Davis) (E), Gold06 (G), SNLS 1yr (S) and SDSS-II (D)) in the context of the Chevalier-Polarski-Linder (CPL) parametrization w(a) = w 0 +w 1 (1−a), according to their Figure of Merit (FoM), their consistency with the cosmological constant (ΛCDM), their consistency with standard rulers (Cosmic Microwave Background (CMB) and Baryon Acoustic Oscillations (BAO)) and their mutual consistency. We find a significant improvement of the FoM (defined as the inverse area of the 95.4% parameter contour) with the number of SnIa of these datasets ((C) highest FoM, (U), (G), (D), (E), (S) lowest FoM). Standard rulers (CMB+BAO) have a better FoM by about a factor of 3, compared to the highest FoM SnIa dataset (C). We also find that the ranking sequence based on consistency with ΛCDM is identical with the corresponding ranking based on consistency with standard rulers ((S) most consistent, (D), (C), (E), (U), (G) least consistent). The ranking sequence of the datasets however changes when we consider the consistency with an expansion history corresponding to evolving dark energy (w 0 ,w 1 ) = (−1.4,2) crossing the phantom divide line w = −1 (it is practically reversed to (G), (U), (E), (S), (D), (C)). The SALT2 and MLCS2k2 fitters are also compared and some peculiar features of the SDSS-II dataset when standardized with the MLCS2k2 fitter are pointed out. Finally, we construct a statistic to estimate the internal consistency of a collection of SnIa datasets. We find that even though there is good consistency among most samples taken from the above datasets, this consistency decreases significantly when the Gold06 (G) dataset is included in the sample

  20. Comparison of Shallow Survey 2012 Multibeam Datasets

    Science.gov (United States)

    Ramirez, T. M.

    2012-12-01

    The purpose of the Shallow Survey common dataset is a comparison of the different technologies utilized for data acquisition in the shallow survey marine environment. The common dataset consists of a series of surveys conducted over a common area of seabed using a variety of systems. It provides equipment manufacturers the opportunity to showcase their latest systems while giving hydrographic researchers and scientists a chance to test their latest algorithms on the dataset so that rigorous comparisons can be made. Five companies collected data for the Common Dataset in the Wellington Harbor area in New Zealand between May 2010 and May 2011; including Kongsberg, Reson, R2Sonic, GeoAcoustics, and Applied Acoustics. The Wellington harbor and surrounding coastal area was selected since it has a number of well-defined features, including the HMNZS South Seas and HMNZS Wellington wrecks, an armored seawall constructed of Tetrapods and Akmons, aquifers, wharves and marinas. The seabed inside the harbor basin is largely fine-grained sediment, with gravel and reefs around the coast. The area outside the harbor on the southern coast is an active environment, with moving sand and exposed reefs. A marine reserve is also in this area. For consistency between datasets, the coastal research vessel R/V Ikatere and crew were used for all surveys conducted for the common dataset. Using Triton's Perspective processing software multibeam datasets collected for the Shallow Survey were processed for detail analysis. Datasets from each sonar manufacturer were processed using the CUBE algorithm developed by the Center for Coastal and Ocean Mapping/Joint Hydrographic Center (CCOM/JHC). Each dataset was gridded at 0.5 and 1.0 meter resolutions for cross comparison and compliance with International Hydrographic Organization (IHO) requirements. Detailed comparisons were made of equipment specifications (transmit frequency, number of beams, beam width), data density, total uncertainty, and

  1. AmeriFlux Network Data Activities: updates, progress and plans

    Science.gov (United States)

    Yang, B.; Boden, T.; Krassovski, M.; Song, X.

    2013-12-01

    The Carbon Dioxide Information Analysis Center (CDIAC) at the Oak Ridge National Laboratory serves as the long-term data repository for the AmeriFlux network. Datasets currently available include hourly or half-hourly meteorological and flux observations, biological measurement records, and synthesis data products. In this presentation, we provide an update of this network database including a comprehensive review and evaluation of the biological data from about 70 sites, development of a new product for flux uncertainty estimates, and re-formatting of Level-2 standard files. In 2013, we also provided data support to two synthesis studies --- 2012 drought synthesis and FACE synthesis. Issues related to data quality and solutions in compiling datasets for these synthesis studies will be discussed. We will also present our work plans in developing and producing other high-level products, such as derivation of phenology from the available measurements at flux sites.

  2. 3DSEM: A 3D microscopy dataset

    Directory of Open Access Journals (Sweden)

    Ahmad P. Tafti

    2016-03-01

    Full Text Available The Scanning Electron Microscope (SEM as a 2D imaging instrument has been widely used in many scientific disciplines including biological, mechanical, and materials sciences to determine the surface attributes of microscopic objects. However the SEM micrographs still remain 2D images. To effectively measure and visualize the surface properties, we need to truly restore the 3D shape model from 2D SEM images. Having 3D surfaces would provide anatomic shape of micro-samples which allows for quantitative measurements and informative visualization of the specimens being investigated. The 3DSEM is a dataset for 3D microscopy vision which is freely available at [1] for any academic, educational, and research purposes. The dataset includes both 2D images and 3D reconstructed surfaces of several real microscopic samples. Keywords: 3D microscopy dataset, 3D microscopy vision, 3D SEM surface reconstruction, Scanning Electron Microscope (SEM

  3. Data Mining for Imbalanced Datasets: An Overview

    Science.gov (United States)

    Chawla, Nitesh V.

    A dataset is imbalanced if the classification categories are not approximately equally represented. Recent years brought increased interest in applying machine learning techniques to difficult "real-world" problems, many of which are characterized by imbalanced data. Additionally the distribution of the testing data may differ from that of the training data, and the true misclassification costs may be unknown at learning time. Predictive accuracy, a popular choice for evaluating performance of a classifier, might not be appropriate when the data is imbalanced and/or the costs of different errors vary markedly. In this Chapter, we discuss some of the sampling techniques used for balancing the datasets, and the performance measures more appropriate for mining imbalanced datasets.

  4. Genomics dataset of unidentified disclosed isolates

    Directory of Open Access Journals (Sweden)

    Bhagwan N. Rekadwad

    2016-09-01

    Full Text Available Analysis of DNA sequences is necessary for higher hierarchical classification of the organisms. It gives clues about the characteristics of organisms and their taxonomic position. This dataset is chosen to find complexities in the unidentified DNA in the disclosed patents. A total of 17 unidentified DNA sequences were thoroughly analyzed. The quick response codes were generated. AT/GC content of the DNA sequences analysis was carried out. The QR is helpful for quick identification of isolates. AT/GC content is helpful for studying their stability at different temperatures. Additionally, a dataset on cleavage code and enzyme code studied under the restriction digestion study, which helpful for performing studies using short DNA sequences was reported. The dataset disclosed here is the new revelatory data for exploration of unique DNA sequences for evaluation, identification, comparison and analysis. Keywords: BioLABs, Blunt ends, Genomics, NEB cutter, Restriction digestion, Short DNA sequences, Sticky ends

  5. Clustering of Emerging Flux

    Science.gov (United States)

    Ruzmaikin, A.

    1997-01-01

    Observations show that newly emerging flux tends to appear on the Solar surface at sites where there is flux already. This results in clustering of solar activity. Standard dynamo theories do not predict this effect.

  6. Harvard Aging Brain Study: Dataset and accessibility.

    Science.gov (United States)

    Dagley, Alexander; LaPoint, Molly; Huijbers, Willem; Hedden, Trey; McLaren, Donald G; Chatwal, Jasmeer P; Papp, Kathryn V; Amariglio, Rebecca E; Blacker, Deborah; Rentz, Dorene M; Johnson, Keith A; Sperling, Reisa A; Schultz, Aaron P

    2017-01-01

    The Harvard Aging Brain Study is sharing its data with the global research community. The longitudinal dataset consists of a 284-subject cohort with the following modalities acquired: demographics, clinical assessment, comprehensive neuropsychological testing, clinical biomarkers, and neuroimaging. To promote more extensive analyses, imaging data was designed to be compatible with other publicly available datasets. A cloud-based system enables access to interested researchers with blinded data available contingent upon completion of a data usage agreement and administrative approval. Data collection is ongoing and currently in its fifth year. Copyright © 2015 Elsevier Inc. All rights reserved.

  7. TAO/TRITON, RAMA, and PIRATA Buoys, Quarterly, 2000-present, Buoyancy Flux

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This dataset has quarterly Buoyancy Flux data from the TAO/TRITON (Pacific Ocean, https://www.pmel.noaa.gov/gtmba/ ), RAMA (Indian Ocean,...

  8. Surface Turbulent Fluxes, 1x1 deg Monthly Climatology, Set1 and NCEP V2c

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c Dataset recently produced through a MEaSURES funded project led by Dr. Chung-Lin Shie...

  9. Surface Turbulent Fluxes, 1x1 deg Yearly Climatology, Set1 and NCEP V2c

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c Dataset recently produced through a MEaSURES funded project led by Dr. Chung-Lin Shie...

  10. TAO/TRITON, RAMA, and PIRATA Buoys, Monthly, 2000-present, Buoyancy Flux

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This dataset has monthly Buoyancy Flux data from the TAO/TRITON (Pacific Ocean, https://www.pmel.noaa.gov/gtmba/ ), RAMA (Indian Ocean,...

  11. Goddard Satellite-Based Surface Turbulent Fluxes Climatology, Yearly Grid V3

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-3 Dataset recently produced through a MEaSUREs funded project led by Dr. Chung-Lin Shie...

  12. Goddard Satellite-Based Surface Turbulent Fluxes Climatology, Seasonal Grid V3

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-3 Dataset recently produced through a MEaSUREs funded project led by Dr. Chung-Lin Shie...

  13. Random Coefficient Logit Model for Large Datasets

    NARCIS (Netherlands)

    C. Hernández-Mireles (Carlos); D. Fok (Dennis)

    2010-01-01

    textabstractWe present an approach for analyzing market shares and products price elasticities based on large datasets containing aggregate sales data for many products, several markets and for relatively long time periods. We consider the recently proposed Bayesian approach of Jiang et al [Jiang,

  14. Thesaurus Dataset of Educational Technology in Chinese

    Science.gov (United States)

    Wu, Linjing; Liu, Qingtang; Zhao, Gang; Huang, Huan; Huang, Tao

    2015-01-01

    The thesaurus dataset of educational technology is a knowledge description of educational technology in Chinese. The aims of this thesaurus were to collect the subject terms in the domain of educational technology, facilitate the standardization of terminology and promote the communication between Chinese researchers and scholars from various…

  15. Evaluation of NASA's Carbon Monitoring System (CMS) Flux Pilot: Terrestrial CO2 Fluxes

    Science.gov (United States)

    Fisher, J. B.; Polhamus, A.; Bowman, K. W.; Collatz, G. J.; Potter, C. S.; Lee, M.; Liu, J.; Jung, M.; Reichstein, M.

    2011-12-01

    NASA's Carbon Monitoring System (CMS) flux pilot project combines NASA's Earth System models in land, ocean and atmosphere to track surface CO2 fluxes. The system is constrained by atmospheric measurements of XCO2 from the Japanese GOSAT satellite, giving a "big picture" view of total CO2 in Earth's atmosphere. Combining two land models (CASA-Ames and CASA-GFED), two ocean models (ECCO2 and NOBM) and two atmospheric chemistry and inversion models (GEOS-5 and GEOS-Chem), the system brings together the stand-alone component models of the Earth System, all of which are run diagnostically constrained by a multitude of other remotely sensed data. Here, we evaluate the biospheric land surface CO2 fluxes (i.e., net ecosystem exchange, NEE) as estimated from the atmospheric flux inversion. We compare against the prior bottom-up estimates (e.g., the CASA models) as well. Our evaluation dataset is the independently derived global wall-to-wall MPI-BGC product, which uses a machine learning algorithm and model tree ensemble to "scale-up" a network of in situ CO2 flux measurements from 253 globally-distributed sites in the FLUXNET network. The measurements are based on the eddy covariance method, which uses observations of co-varying fluxes of CO2 (and water and energy) from instruments on towers extending above ecosystem canopies; the towers integrate fluxes over large spatial areas (~1 km2). We present global maps of CO2 fluxes and differences between products, summaries of fluxes by TRANSCOM region, country, latitude, and biome type, and assess the time series, including timing of minimum and maximum fluxes. This evaluation shows both where the CMS is performing well, and where improvements should be directed in further work.

  16. Sharing Video Datasets in Design Research

    DEFF Research Database (Denmark)

    Christensen, Bo; Abildgaard, Sille Julie Jøhnk

    2017-01-01

    This paper examines how design researchers, design practitioners and design education can benefit from sharing a dataset. We present the Design Thinking Research Symposium 11 (DTRS11) as an exemplary project that implied sharing video data of design processes and design activity in natural settings...... with a large group of fellow academics from the international community of Design Thinking Research, for the purpose of facilitating research collaboration and communication within the field of Design and Design Thinking. This approach emphasizes the social and collaborative aspects of design research, where...... a multitude of appropriate perspectives and methods may be utilized in analyzing and discussing the singular dataset. The shared data is, from this perspective, understood as a design object in itself, which facilitates new ways of working, collaborating, studying, learning and educating within the expanding...

  17. Automatic processing of multimodal tomography datasets.

    Science.gov (United States)

    Parsons, Aaron D; Price, Stephen W T; Wadeson, Nicola; Basham, Mark; Beale, Andrew M; Ashton, Alun W; Mosselmans, J Frederick W; Quinn, Paul D

    2017-01-01

    With the development of fourth-generation high-brightness synchrotrons on the horizon, the already large volume of data that will be collected on imaging and mapping beamlines is set to increase by orders of magnitude. As such, an easy and accessible way of dealing with such large datasets as quickly as possible is required in order to be able to address the core scientific problems during the experimental data collection. Savu is an accessible and flexible big data processing framework that is able to deal with both the variety and the volume of data of multimodal and multidimensional scientific datasets output such as those from chemical tomography experiments on the I18 microfocus scanning beamline at Diamond Light Source.

  18. Interpolation of diffusion weighted imaging datasets

    DEFF Research Database (Denmark)

    Dyrby, Tim B; Lundell, Henrik; Burke, Mark W

    2014-01-01

    anatomical details and signal-to-noise-ratio for reliable fibre reconstruction. We assessed the potential benefits of interpolating DWI datasets to a higher image resolution before fibre reconstruction using a diffusion tensor model. Simulations of straight and curved crossing tracts smaller than or equal......Diffusion weighted imaging (DWI) is used to study white-matter fibre organisation, orientation and structural connectivity by means of fibre reconstruction algorithms and tractography. For clinical settings, limited scan time compromises the possibilities to achieve high image resolution for finer...... interpolation methods fail to disentangle fine anatomical details if PVE is too pronounced in the original data. As for validation we used ex-vivo DWI datasets acquired at various image resolutions as well as Nissl-stained sections. Increasing the image resolution by a factor of eight yielded finer geometrical...

  19. Data assimilation and model evaluation experiment datasets

    Science.gov (United States)

    Lai, Chung-Cheng A.; Qian, Wen; Glenn, Scott M.

    1994-01-01

    The Institute for Naval Oceanography, in cooperation with Naval Research Laboratories and universities, executed the Data Assimilation and Model Evaluation Experiment (DAMEE) for the Gulf Stream region during fiscal years 1991-1993. Enormous effort has gone into the preparation of several high-quality and consistent datasets for model initialization and verification. This paper describes the preparation process, the temporal and spatial scopes, the contents, the structure, etc., of these datasets. The goal of DAMEE and the need of data for the four phases of experiment are briefly stated. The preparation of DAMEE datasets consisted of a series of processes: (1) collection of observational data; (2) analysis and interpretation; (3) interpolation using the Optimum Thermal Interpolation System package; (4) quality control and re-analysis; and (5) data archiving and software documentation. The data products from these processes included a time series of 3D fields of temperature and salinity, 2D fields of surface dynamic height and mixed-layer depth, analysis of the Gulf Stream and rings system, and bathythermograph profiles. To date, these are the most detailed and high-quality data for mesoscale ocean modeling, data assimilation, and forecasting research. Feedback from ocean modeling groups who tested this data was incorporated into its refinement. Suggestions for DAMEE data usages include (1) ocean modeling and data assimilation studies, (2) diagnosis and theoretical studies, and (3) comparisons with locally detailed observations.

  20. A hybrid organic-inorganic perovskite dataset

    Science.gov (United States)

    Kim, Chiho; Huan, Tran Doan; Krishnan, Sridevi; Ramprasad, Rampi

    2017-05-01

    Hybrid organic-inorganic perovskites (HOIPs) have been attracting a great deal of attention due to their versatility of electronic properties and fabrication methods. We prepare a dataset of 1,346 HOIPs, which features 16 organic cations, 3 group-IV cations and 4 halide anions. Using a combination of an atomic structure search method and density functional theory calculations, the optimized structures, the bandgap, the dielectric constant, and the relative energies of the HOIPs are uniformly prepared and validated by comparing with relevant experimental and/or theoretical data. We make the dataset available at Dryad Digital Repository, NoMaD Repository, and Khazana Repository (http://khazana.uconn.edu/), hoping that it could be useful for future data-mining efforts that can explore possible structure-property relationships and phenomenological models. Progressive extension of the dataset is expected as new organic cations become appropriate within the HOIP framework, and as additional properties are calculated for the new compounds found.

  1. Compact neutron flux monitor

    International Nuclear Information System (INIS)

    Madhavi, V.; Phatak, P.R.; Bahadur, C.; Bayala, A.K.; Jakati, R.K.; Sathian, V.

    2003-01-01

    Full text: A compact size neutron flux monitor has been developed incorporating standard boards developed for smart radiation monitors. The sensitivity of the monitors is 0.4cps/nV. It has been tested up to 2075 nV flux with standard neutron sources. It shows convincing results even in high flux areas like 6m away from the accelerator in RMC (Parel) for 106/107 nV. These monitors have a focal and remote display, alarm function with potential free contacts for centralized control and additional provision of connectivity via RS485/Ethernet. This paper describes the construction, working and results of the above flux monitor

  2. Quantifying uncertainty in observational rainfall datasets

    Science.gov (United States)

    Lennard, Chris; Dosio, Alessandro; Nikulin, Grigory; Pinto, Izidine; Seid, Hussen

    2015-04-01

    The CO-ordinated Regional Downscaling Experiment (CORDEX) has to date seen the publication of at least ten journal papers that examine the African domain during 2012 and 2013. Five of these papers consider Africa generally (Nikulin et al. 2012, Kim et al. 2013, Hernandes-Dias et al. 2013, Laprise et al. 2013, Panitz et al. 2013) and five have regional foci: Tramblay et al. (2013) on Northern Africa, Mariotti et al. (2014) and Gbobaniyi el al. (2013) on West Africa, Endris et al. (2013) on East Africa and Kalagnoumou et al. (2013) on southern Africa. There also are a further three papers that the authors know about under review. These papers all use an observed rainfall and/or temperature data to evaluate/validate the regional model output and often proceed to assess projected changes in these variables due to climate change in the context of these observations. The most popular reference rainfall data used are the CRU, GPCP, GPCC, TRMM and UDEL datasets. However, as Kalagnoumou et al. (2013) point out there are many other rainfall datasets available for consideration, for example, CMORPH, FEWS, TAMSAT & RIANNAA, TAMORA and the WATCH & WATCH-DEI data. They, with others (Nikulin et al. 2012, Sylla et al. 2012) show that the observed datasets can have a very wide spread at a particular space-time coordinate. As more ground, space and reanalysis-based rainfall products become available, all which use different methods to produce precipitation data, the selection of reference data is becoming an important factor in model evaluation. A number of factors can contribute to a uncertainty in terms of the reliability and validity of the datasets such as radiance conversion algorithims, the quantity and quality of available station data, interpolation techniques and blending methods used to combine satellite and guage based products. However, to date no comprehensive study has been performed to evaluate the uncertainty in these observational datasets. We assess 18 gridded

  3. Primary cosmic ray flux

    Energy Technology Data Exchange (ETDEWEB)

    Stanev, Todor

    2001-05-01

    We discuss the primary cosmic ray flux from the point of view of particle interactions and production of atmospheric neutrinos. The overall normalization of the cosmic ray flux and its time variations and site dependence are major ingredients of the atmospheric neutrino predictions and the basis for the derivation of the neutrino oscillation parameters.

  4. Flux cutting in superconductors

    International Nuclear Information System (INIS)

    Campbell, A M

    2011-01-01

    This paper describes experiments and theories of flux cutting in superconductors. The use of the flux line picture in free space is discussed. In superconductors cutting can either be by means of flux at an angle to other layers of flux, as in longitudinal current experiments, or due to shearing of the vortex lattice as in grain boundaries in YBCO. Experiments on longitudinal currents can be interpreted in terms of flux rings penetrating axial lines. More physical models of flux cutting are discussed but all predict much larger flux cutting forces than are observed. Also, cutting is occurring at angles between vortices of about one millidegree which is hard to explain. The double critical state model and its developments are discussed in relation to experiments on crossed and rotating fields. A new experiment suggested by Clem gives more direct information. It shows that an elliptical yield surface of the critical state works well, but none of the theoretical proposals for determining the direction of E are universally applicable. It appears that, as soon as any flux flow takes place, cutting also occurs. The conclusion is that new theories are required. (perspective)

  5. Development of a SPARK Training Dataset

    Energy Technology Data Exchange (ETDEWEB)

    Sayre, Amanda M. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Olson, Jarrod R. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2015-03-01

    In its first five years, the National Nuclear Security Administration’s (NNSA) Next Generation Safeguards Initiative (NGSI) sponsored more than 400 undergraduate, graduate, and post-doctoral students in internships and research positions (Wyse 2012). In the past seven years, the NGSI program has, and continues to produce a large body of scientific, technical, and policy work in targeted core safeguards capabilities and human capital development activities. Not only does the NGSI program carry out activities across multiple disciplines, but also across all U.S. Department of Energy (DOE)/NNSA locations in the United States. However, products are not readily shared among disciplines and across locations, nor are they archived in a comprehensive library. Rather, knowledge of NGSI-produced literature is localized to the researchers, clients, and internal laboratory/facility publication systems such as the Electronic Records and Information Capture Architecture (ERICA) at the Pacific Northwest National Laboratory (PNNL). There is also no incorporated way of analyzing existing NGSI literature to determine whether the larger NGSI program is achieving its core safeguards capabilities and activities. A complete library of NGSI literature could prove beneficial to a cohesive, sustainable, and more economical NGSI program. The Safeguards Platform for Automated Retrieval of Knowledge (SPARK) has been developed to be a knowledge storage, retrieval, and analysis capability to capture safeguards knowledge to exist beyond the lifespan of NGSI. During the development process, it was necessary to build a SPARK training dataset (a corpus of documents) for initial entry into the system and for demonstration purposes. We manipulated these data to gain new information about the breadth of NGSI publications, and they evaluated the science-policy interface at PNNL as a practical demonstration of SPARK’s intended analysis capability. The analysis demonstration sought to answer the

  6. Development of a SPARK Training Dataset

    International Nuclear Information System (INIS)

    Sayre, Amanda M.; Olson, Jarrod R.

    2015-01-01

    In its first five years, the National Nuclear Security Administration's (NNSA) Next Generation Safeguards Initiative (NGSI) sponsored more than 400 undergraduate, graduate, and post-doctoral students in internships and research positions (Wyse 2012). In the past seven years, the NGSI program has, and continues to produce a large body of scientific, technical, and policy work in targeted core safeguards capabilities and human capital development activities. Not only does the NGSI program carry out activities across multiple disciplines, but also across all U.S. Department of Energy (DOE)/NNSA locations in the United States. However, products are not readily shared among disciplines and across locations, nor are they archived in a comprehensive library. Rather, knowledge of NGSI-produced literature is localized to the researchers, clients, and internal laboratory/facility publication systems such as the Electronic Records and Information Capture Architecture (ERICA) at the Pacific Northwest National Laboratory (PNNL). There is also no incorporated way of analyzing existing NGSI literature to determine whether the larger NGSI program is achieving its core safeguards capabilities and activities. A complete library of NGSI literature could prove beneficial to a cohesive, sustainable, and more economical NGSI program. The Safeguards Platform for Automated Retrieval of Knowledge (SPARK) has been developed to be a knowledge storage, retrieval, and analysis capability to capture safeguards knowledge to exist beyond the lifespan of NGSI. During the development process, it was necessary to build a SPARK training dataset (a corpus of documents) for initial entry into the system and for demonstration purposes. We manipulated these data to gain new information about the breadth of NGSI publications, and they evaluated the science-policy interface at PNNL as a practical demonstration of SPARK's intended analysis capability. The analysis demonstration sought to answer

  7. Heat flux microsensor measurements

    Science.gov (United States)

    Terrell, J. P.; Hager, J. M.; Onishi, S.; Diller, T. E.

    1992-01-01

    A thin-film heat flux sensor has been fabricated on a stainless steel substrate. The thermocouple elements of the heat flux sensor were nickel and nichrome, and the temperature resistance sensor was platinum. The completed heat flux microsensor was calibrated at the AEDC radiation facility. The gage output was linear with heat flux with no apparent temperature effect on sensitivity. The gage was used for heat flux measurements at the NASA Langley Vitiated Air Test Facility. Vitiated air was expanded to Mach 3.0 and hydrogen fuel was injected. Measurements were made on the wall of a diverging duct downstream of the injector during all stages of the hydrogen combustion tests. Because the wall and the gage were not actively cooled, the wall temperature reached over 1000 C (1900 F) during the most severe test.

  8. Developing a Data-Set for Stereopsis

    Directory of Open Access Journals (Sweden)

    D.W Hunter

    2014-08-01

    Full Text Available Current research on binocular stereopsis in humans and non-human primates has been limited by a lack of available data-sets. Current data-sets fall into two categories; stereo-image sets with vergence but no ranging information (Hibbard, 2008, Vision Research, 48(12, 1427-1439 or combinations of depth information with binocular images and video taken from cameras in fixed fronto-parallel configurations exhibiting neither vergence or focus effects (Hirschmuller & Scharstein, 2007, IEEE Conf. Computer Vision and Pattern Recognition. The techniques for generating depth information are also imperfect. Depth information is normally inaccurate or simply missing near edges and on partially occluded surfaces. For many areas of vision research these are the most interesting parts of the image (Goutcher, Hunter, Hibbard, 2013, i-Perception, 4(7, 484; Scarfe & Hibbard, 2013, Vision Research. Using state-of-the-art open-source ray-tracing software (PBRT as a back-end, our intention is to release a set of tools that will allow researchers in this field to generate artificial binocular stereoscopic data-sets. Although not as realistic as photographs, computer generated images have significant advantages in terms of control over the final output and ground-truth information about scene depth is easily calculated at all points in the scene, even partially occluded areas. While individual researchers have been developing similar stimuli by hand for many decades, we hope that our software will greatly reduce the time and difficulty of creating naturalistic binocular stimuli. Our intension in making this presentation is to elicit feedback from the vision community about what sort of features would be desirable in such software.

  9. Overview of the 2013 FireFlux II grass fire field experiment

    Science.gov (United States)

    C.B. Clements; B. Davis; D. Seto; J. Contezac; A. Kochanski; J.-B. Fillipi; N. Lareau; B. Barboni; B. Butler; S. Krueger; R. Ottmar; R. Vihnanek; W.E. Heilman; J. Flynn; M.A. Jenkins; J. Mandel; C. Teske; D. Jimenez; J. O' Brien; B. Lefer

    2014-01-01

    In order to better understand the dynamics of fire-atmosphere interactions and the role of micrometeorology on fire behaviour the FireFlux campaign was conducted in 2006 on a coastal tall-grass prairie in southeast Texas, USA. The FireFlux campaign dataset has become the international standard for evaluating coupled fire-atmosphere model systems. While FireFlux is one...

  10. Quality Controlling CMIP datasets at GFDL

    Science.gov (United States)

    Horowitz, L. W.; Radhakrishnan, A.; Balaji, V.; Adcroft, A.; Krasting, J. P.; Nikonov, S.; Mason, E. E.; Schweitzer, R.; Nadeau, D.

    2017-12-01

    As GFDL makes the switch from model development to production in light of the Climate Model Intercomparison Project (CMIP), GFDL's efforts are shifted to testing and more importantly establishing guidelines and protocols for Quality Controlling and semi-automated data publishing. Every CMIP cycle introduces key challenges and the upcoming CMIP6 is no exception. The new CMIP experimental design comprises of multiple MIPs facilitating research in different focus areas. This paradigm has implications not only for the groups that develop the models and conduct the runs, but also for the groups that monitor, analyze and quality control the datasets before data publishing, before their knowledge makes its way into reports like the IPCC (Intergovernmental Panel on Climate Change) Assessment Reports. In this talk, we discuss some of the paths taken at GFDL to quality control the CMIP-ready datasets including: Jupyter notebooks, PrePARE, LAMP (Linux, Apache, MySQL, PHP/Python/Perl): technology-driven tracker system to monitor the status of experiments qualitatively and quantitatively, provide additional metadata and analysis services along with some in-built controlled-vocabulary validations in the workflow. In addition to this, we also discuss the integration of community-based model evaluation software (ESMValTool, PCMDI Metrics Package, and ILAMB) as part of our CMIP6 workflow.

  11. Integrated remotely sensed datasets for disaster management

    Science.gov (United States)

    McCarthy, Timothy; Farrell, Ronan; Curtis, Andrew; Fotheringham, A. Stewart

    2008-10-01

    Video imagery can be acquired from aerial, terrestrial and marine based platforms and has been exploited for a range of remote sensing applications over the past two decades. Examples include coastal surveys using aerial video, routecorridor infrastructures surveys using vehicle mounted video cameras, aerial surveys over forestry and agriculture, underwater habitat mapping and disaster management. Many of these video systems are based on interlaced, television standards such as North America's NTSC and European SECAM and PAL television systems that are then recorded using various video formats. This technology has recently being employed as a front-line, remote sensing technology for damage assessment post-disaster. This paper traces the development of spatial video as a remote sensing tool from the early 1980s to the present day. The background to a new spatial-video research initiative based at National University of Ireland, Maynooth, (NUIM) is described. New improvements are proposed and include; low-cost encoders, easy to use software decoders, timing issues and interoperability. These developments will enable specialists and non-specialists collect, process and integrate these datasets within minimal support. This integrated approach will enable decision makers to access relevant remotely sensed datasets quickly and so, carry out rapid damage assessment during and post-disaster.

  12. AmeriFlux Data System Looking Forward

    Science.gov (United States)

    Agarwal, D.; Cheah, Y. W.; Chu, H.; Keenan, T. F.; Pastorello, G.; Torn, M. S.; Humphrey, M.; Beekwilder, N.

    2016-12-01

    The AmeriFlux data system is turning 20 along with the network. This anniversary provides an opportunity to reimagine the data system and services to support the next 20 years of AmeriFlux. The next generation goals of the data system will be continuous data quality and consistency improvements, and also increased services to both tower owners and data users. An example change is new data citation capabilities and support leveraging Digital Object Identifiers (DOI) to enable easier citation and improved tracking of data usage. A new processing pipeline will enable more regular production of gap-filled and partitioned datasets. Automation will be continuously increased to help improve the time from data submission to user data availability. New standard data submission and processing formats will enable automation of the data submission and improved QA/QC processes. This new data system is made possible by the extensive work that the AmeriFlux data team at CDIAC has done over the last 20 years in close collaboration with the tower teams. This presentation will provide both a historical perspective and a vision for the future of the AmeriFlux data system.

  13. Surface Turbulent Fluxes, 1x1 deg Yearly Climatology, Set1 and NCEP V2c (GSSTFYC) at GES DISC

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c Dataset recently produced through a MEaSURES funded project led by Dr. Chung-Lin Shie...

  14. Surface Turbulent Fluxes, 1x1 deg Monthly Climatology, Set1 and NCEP V2c (GSSTFMC) at GES DISC

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c Dataset recently produced through a MEaSURES funded project led by Dr. Chung-Lin Shie...

  15. Surface Turbulent Fluxes, 1x1 deg Monthly Grid, Set1 and Interpolated Data V2c

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c Dataset recently produced through a MEaSUREs funded project led by Dr. Chung-Lin Shie...

  16. Carbon Monitoring System Flux for Fossil Fuel L4 V1 (CMSFluxFossilfuel) at GES DISC

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset provides the Carbon Flux for Fossil Fuel. The NASA Carbon Monitoring System (CMS) is designed to make significant contributions in characterizing,...

  17. Quantifying benthic nitrogen fluxes in Puget Sound, Washington: a review of available data

    Science.gov (United States)

    Sheibley, Richard W.; Paulson, Anthony J.

    2014-01-01

    Understanding benthic fluxes is important for understanding the fate of materials that settle to the Puget Sound, Washington, seafloor, as well as the impact these fluxes have on the chemical composition and biogeochemical cycles of marine waters. Existing approaches used to measure benthic nitrogen flux in Puget Sound and elsewhere were reviewed and summarized, and factors for considering each approach were evaluated. Factors for selecting an appropriate approach for gathering information about benthic flux include: availability of resources, objectives of projects, and determination of which processes each approach measures. An extensive search of literature was undertaken to summarize known benthic nitrogen fluxes in Puget Sound. A total of 138 individual flux chamber measurements and 38 sets of diffusive fluxes were compiled for this study. Of the diffusive fluxes, 35 new datasets were located, and new flux calculations are presented in this report. About 65 new diffusive flux calculations are provided across all nitrogen species (nitrate, NO3-; nitrite, NO2-; ammonium, NH4+). Data analysis of this newly compiled benthic flux dataset showed that fluxes beneath deep (greater than 50 meters) water tended to be lower than those beneath shallow (less than 50 meters) water. Additionally, variability in flux at the shallow depths was greater, possibly indicating a more dynamic interaction between the benthic and pelagic environments. The overall range of bottom temperatures from studies in the Puget Sound area were small (5–16 degrees Celsius), and only NH4+ flux showed any pattern with temperature. For NH4+, flux values and variability increased at greater than about 12 degrees Celsius. Collection of additional study site metadata about environmental factors (bottom temperature, depth, sediment porosity, sediment type, and sediment organic matter) will help with development of a broader regional understanding benthic nitrogen flux in the Puget Sound.

  18. Strontium removal jar test dataset for all figures and tables.

    Data.gov (United States)

    U.S. Environmental Protection Agency — The datasets where used to generate data to demonstrate strontium removal under various water quality and treatment conditions. This dataset is associated with the...

  19. Continuous magnetic flux pump

    Science.gov (United States)

    Hildebrandt, A. F.; Elleman, D. D.; Whitmore, F. C. (Inventor)

    1966-01-01

    A method and means for altering the intensity of a magnetic field by transposing flux from one location to the location desired fro the magnetic field are examined. The device described includes a pair of communicating cavities formed in a block of superconducting material, is dimensioned to be insertable into one of the cavities and to substantially fill the cavity. Magnetic flux is first trapped in the cavities by establishing a magnetic field while the superconducting material is above the critical temperature at which it goes superconducting. Thereafter, the temperature of the material is reduced below the critical value, and then the exciting magnetic field may be removed. By varying the ratios of the areas of the two cavities, it is possible to produce a field having much greater flux density in the second, smaller cavity, into which the flux transposed.

  20. Flux in Tallinn

    Index Scriptorium Estoniae

    2004-01-01

    Rahvusvahelise elektroonilise kunsti sümpoosioni ISEA2004 klubiõhtu "Flux in Tallinn" klubis Bon Bon. Eestit esindasid Ropotator, Ars Intel Inc., Urmas Puhkan, Joel Tammik, Taavi Tulev (pseud. Wochtzchee). Klubiõhtu koordinaator Andres Lõo

  1. Flux shunts for undulators

    International Nuclear Information System (INIS)

    Hoyer, E.; Chin, J.; Hassenzahl, W.V.

    1993-05-01

    Undulators for high-performance applications in synchrotron-radiation sources and periodic magnetic structures for free-electron lasers have stringent requirements on the curvature of the electron's average trajectory. Undulators using the permanent magnet hybrid configuration often have fields in their central region that produce a curved trajectory caused by local, ambient magnetic fields such as those of the earth. The 4.6 m long Advanced Light Source (ALS) undulators use flux shunts to reduce this effect. These flux shunts are magnetic linkages of very high permeability material connecting the two steel beams that support the magnetic structures. The shunts reduce the scalar potential difference between the supporting beams and carry substantial flux that would normally appear in the undulator gap. Magnetic design, mechanical configuration of the flux shunts and magnetic measurements of their effect on the ALS undulators are described

  2. Predicting dataset popularity for the CMS experiment

    CERN Document Server

    INSPIRE-00005122; Li, Ting; Giommi, Luca; Bonacorsi, Daniele; Wildish, Tony

    2016-01-01

    The CMS experiment at the LHC accelerator at CERN relies on its computing infrastructure to stay at the frontier of High Energy Physics, searching for new phenomena and making discoveries. Even though computing plays a significant role in physics analysis we rarely use its data to predict the system behavior itself. A basic information about computing resources, user activities and site utilization can be really useful for improving the throughput of the system and its management. In this paper, we discuss a first CMS analysis of dataset popularity based on CMS meta-data which can be used as a model for dynamic data placement and provide the foundation of data-driven approach for the CMS computing infrastructure.

  3. Predicting dataset popularity for the CMS experiment

    International Nuclear Information System (INIS)

    Kuznetsov, V.; Li, T.; Giommi, L.; Bonacorsi, D.; Wildish, T.

    2016-01-01

    The CMS experiment at the LHC accelerator at CERN relies on its computing infrastructure to stay at the frontier of High Energy Physics, searching for new phenomena and making discoveries. Even though computing plays a significant role in physics analysis we rarely use its data to predict the system behavior itself. A basic information about computing resources, user activities and site utilization can be really useful for improving the throughput of the system and its management. In this paper, we discuss a first CMS analysis of dataset popularity based on CMS meta-data which can be used as a model for dynamic data placement and provide the foundation of data-driven approach for the CMS computing infrastructure. (paper)

  4. Internationally coordinated glacier monitoring: strategy and datasets

    Science.gov (United States)

    Hoelzle, Martin; Armstrong, Richard; Fetterer, Florence; Gärtner-Roer, Isabelle; Haeberli, Wilfried; Kääb, Andreas; Kargel, Jeff; Nussbaumer, Samuel; Paul, Frank; Raup, Bruce; Zemp, Michael

    2014-05-01

    (c) the Randolph Glacier Inventory (RGI), a new and globally complete digital dataset of outlines from about 180,000 glaciers with some meta-information, which has been used for many applications relating to the IPCC AR5 report. Concerning glacier changes, a database (Fluctuations of Glaciers) exists containing information about mass balance, front variations including past reconstructed time series, geodetic changes and special events. Annual mass balance reporting contains information for about 125 glaciers with a subset of 37 glaciers with continuous observational series since 1980 or earlier. Front variation observations of around 1800 glaciers are available from most of the mountain ranges world-wide. This database was recently updated with 26 glaciers having an unprecedented dataset of length changes from from reconstructions of well-dated historical evidence going back as far as the 16th century. Geodetic observations of about 430 glaciers are available. The database is completed by a dataset containing information on special events including glacier surges, glacier lake outbursts, ice avalanches, eruptions of ice-clad volcanoes, etc. related to about 200 glaciers. A special database of glacier photographs contains 13,000 pictures from around 500 glaciers, some of them dating back to the 19th century. A key challenge is to combine and extend the traditional observations with fast evolving datasets from new technologies.

  5. MIPS bacterial genomes functional annotation benchmark dataset.

    Science.gov (United States)

    Tetko, Igor V; Brauner, Barbara; Dunger-Kaltenbach, Irmtraud; Frishman, Goar; Montrone, Corinna; Fobo, Gisela; Ruepp, Andreas; Antonov, Alexey V; Surmeli, Dimitrij; Mewes, Hans-Wernen

    2005-05-15

    Any development of new methods for automatic functional annotation of proteins according to their sequences requires high-quality data (as benchmark) as well as tedious preparatory work to generate sequence parameters required as input data for the machine learning methods. Different program settings and incompatible protocols make a comparison of the analyzed methods difficult. The MIPS Bacterial Functional Annotation Benchmark dataset (MIPS-BFAB) is a new, high-quality resource comprising four bacterial genomes manually annotated according to the MIPS functional catalogue (FunCat). These resources include precalculated sequence parameters, such as sequence similarity scores, InterPro domain composition and other parameters that could be used to develop and benchmark methods for functional annotation of bacterial protein sequences. These data are provided in XML format and can be used by scientists who are not necessarily experts in genome annotation. BFAB is available at http://mips.gsf.de/proj/bfab

  6. 2006 Fynmeet sea clutter measurement trial: Datasets

    CSIR Research Space (South Africa)

    Herselman, PLR

    2007-09-06

    Full Text Available -011............................................................................................................................................................................................. 25 iii Dataset CAD14-001 0 5 10 15 20 25 30 35 10 20 30 40 50 60 70 80 90 R an ge G at e # Time [s] A bs ol ut e R an ge [m ] RCS [dBm2] vs. time and range for f1 = 9.000 GHz - CAD14-001 2400 2600 2800... 40 10 20 30 40 50 60 70 80 90 R an ge G at e # Time [s] A bs ol ut e R an ge [m ] RCS [dBm2] vs. time and range for f1 = 9.000 GHz - CAD14-002 2400 2600 2800 3000 3200 3400 3600 -30 -25 -20 -15 -10 -5 0 5 10...

  7. A new bed elevation dataset for Greenland

    Directory of Open Access Journals (Sweden)

    J. L. Bamber

    2013-03-01

    Full Text Available We present a new bed elevation dataset for Greenland derived from a combination of multiple airborne ice thickness surveys undertaken between the 1970s and 2012. Around 420 000 line kilometres of airborne data were used, with roughly 70% of this having been collected since the year 2000, when the last comprehensive compilation was undertaken. The airborne data were combined with satellite-derived elevations for non-glaciated terrain to produce a consistent bed digital elevation model (DEM over the entire island including across the glaciated–ice free boundary. The DEM was extended to the continental margin with the aid of bathymetric data, primarily from a compilation for the Arctic. Ice thickness was determined where an ice shelf exists from a combination of surface elevation and radar soundings. The across-track spacing between flight lines warranted interpolation at 1 km postings for significant sectors of the ice sheet. Grids of ice surface elevation, error estimates for the DEM, ice thickness and data sampling density were also produced alongside a mask of land/ocean/grounded ice/floating ice. Errors in bed elevation range from a minimum of ±10 m to about ±300 m, as a function of distance from an observation and local topographic variability. A comparison with the compilation published in 2001 highlights the improvement in resolution afforded by the new datasets, particularly along the ice sheet margin, where ice velocity is highest and changes in ice dynamics most marked. We estimate that the volume of ice included in our land-ice mask would raise mean sea level by 7.36 m, excluding any solid earth effects that would take place during ice sheet decay.

  8. Neutron flux monitor

    International Nuclear Information System (INIS)

    Oda, Naotaka.

    1993-01-01

    The device of the present invention greatly saves an analog processing section such as an analog filter and an analog processing circuit. That is, the device of the present invention comprises (1) a neutron flux detection means for detecting neutron fluxed in the reactor, (2) a digital filter means for dividing signals corresponding to the detected neutron fluxes into predetermined frequency band regions, (3) a calculation processing means for applying a calculation processing corresponding to the frequency band regions to the neutron flux detection signals divided by the digital filter means. With such a constitution, since the neutron detection signals are processed by the digital filter means, the accuracy is improved and the change for the property of the filter is facilitated. Further, when a neutron flux level is obtained, a calculation processing corresponding to the frequency band region can be conducted without the analog processing circuit. Accordingly, maintenance and accuracy are improved by greatly decreasing the number of parts. Further, since problems inherent to the analog circuit are solved, neutron fluxes are monitored at high reliability. (I.S.)

  9. Neutron flux monitoring device

    International Nuclear Information System (INIS)

    Shimazu, Yoichiro.

    1995-01-01

    In a neutron flux monitoring device, there are disposed a neutron flux measuring means for outputting signals in accordance with the intensity of neutron fluxes, a calculation means for calculating a self power density spectrum at a frequency band suitable to an object to be measured based on the output of the neutron flux measuring means, an alarm set value generation means for outputting an alarm set value as a comparative reference, and an alarm judging means for comparing the alarm set value with the outputted value of the calculation means to judge requirement of generating an alarm and generate an alarm in accordance with the result of the judgement. Namely, the time-series of neutron flux signals is put to fourier transformation for a predetermined period of time by the calculation means, and from each of square sums for real number component and imaginary number component for each of the frequencies, a self power density spectrum in the frequency band suitable to the object to be measured is calculated. Then, when the set reference value is exceeded, an alarm is generated. This can reliably prevent generation of erroneous alarm due to neutron flux noises and can accurately generate an alarm at an appropriate time. (N.H.)

  10. Wind Integration National Dataset Toolkit | Grid Modernization | NREL

    Science.gov (United States)

    Integration National Dataset Toolkit Wind Integration National Dataset Toolkit The Wind Integration National Dataset (WIND) Toolkit is an update and expansion of the Eastern Wind Integration Data Set and Western Wind Integration Data Set. It supports the next generation of wind integration studies. WIND

  11. Solar Integration National Dataset Toolkit | Grid Modernization | NREL

    Science.gov (United States)

    Solar Integration National Dataset Toolkit Solar Integration National Dataset Toolkit NREL is working on a Solar Integration National Dataset (SIND) Toolkit to enable researchers to perform U.S . regional solar generation integration studies. It will provide modeled, coherent subhourly solar power data

  12. Technical note: An inorganic water chemistry dataset (1972–2011 ...

    African Journals Online (AJOL)

    A national dataset of inorganic chemical data of surface waters (rivers, lakes, and dams) in South Africa is presented and made freely available. The dataset comprises more than 500 000 complete water analyses from 1972 up to 2011, collected from more than 2 000 sample monitoring stations in South Africa. The dataset ...

  13. QSAR ligand dataset for modelling mutagenicity, genotoxicity, and rodent carcinogenicity

    Directory of Open Access Journals (Sweden)

    Davy Guan

    2018-04-01

    Full Text Available Five datasets were constructed from ligand and bioassay result data from the literature. These datasets include bioassay results from the Ames mutagenicity assay, Greenscreen GADD-45a-GFP assay, Syrian Hamster Embryo (SHE assay, and 2 year rat carcinogenicity assay results. These datasets provide information about chemical mutagenicity, genotoxicity and carcinogenicity.

  14. Hydrology Research with the North American Land Data Assimilation System (NLDAS) Datasets at the NASA GES DISC Using Giovanni

    Science.gov (United States)

    Mocko, David M.; Rui, Hualan; Acker, James G.

    2013-01-01

    The North American Land Data Assimilation System (NLDAS) is a collaboration project between NASA/GSFC, NOAA, Princeton Univ., and the Univ. of Washington. NLDAS has created a surface meteorology dataset using the best-available observations and reanalyses the backbone of this dataset is a gridded precipitation analysis from rain gauges. This dataset is used to drive four separate land-surface models (LSMs) to produce datasets of soil moisture, snow, runoff, and surface fluxes. NLDAS datasets are available hourly and extend from Jan 1979 to near real-time with a typical 4-day lag. The datasets are available at 1/8th-degree over CONUS and portions of Canada and Mexico from 25-53 North. The datasets have been extensively evaluated against observations, and are also used as part of a drought monitor. NLDAS datasets are available from the NASA GES DISC and can be accessed via ftp, GDS, Mirador, and Giovanni. GES DISC news articles were published showing figures from the heat wave of 2011, Hurricane Irene, Tropical Storm Lee, and the low-snow winter of 2011-2012. For this presentation, Giovanni-generated figures using NLDAS data from the derecho across the U.S. Midwest and Mid-Atlantic will be presented. Also, similar figures will be presented from the landfall of Hurricane Isaac and the before-and-after drought conditions of the path of the tropical moisture into the central states of the U.S. Updates on future products and datasets from the NLDAS project will also be introduced.

  15. The Open Flux Problem

    Science.gov (United States)

    Linker, J. A.; Caplan, R. M.; Downs, C.; Riley, P.; Mikic, Z.; Lionello, R.; Henney, C. J.; Arge, C. N.; Liu, Y.; Derosa, M. L.; Yeates, A.; Owens, M. J.

    2017-10-01

    The heliospheric magnetic field is of pivotal importance in solar and space physics. The field is rooted in the Sun’s photosphere, where it has been observed for many years. Global maps of the solar magnetic field based on full-disk magnetograms are commonly used as boundary conditions for coronal and solar wind models. Two primary observational constraints on the models are (1) the open field regions in the model should approximately correspond to coronal holes (CHs) observed in emission and (2) the magnitude of the open magnetic flux in the model should match that inferred from in situ spacecraft measurements. In this study, we calculate both magnetohydrodynamic and potential field source surface solutions using 14 different magnetic maps produced from five different types of observatory magnetograms, for the time period surrounding 2010 July. We have found that for all of the model/map combinations, models that have CH areas close to observations underestimate the interplanetary magnetic flux, or, conversely, for models to match the interplanetary flux, the modeled open field regions are larger than CHs observed in EUV emission. In an alternative approach, we estimate the open magnetic flux entirely from solar observations by combining automatically detected CHs for Carrington rotation 2098 with observatory synoptic magnetic maps. This approach also underestimates the interplanetary magnetic flux. Our results imply that either typical observatory maps underestimate the Sun’s magnetic flux, or a significant portion of the open magnetic flux is not rooted in regions that are obviously dark in EUV and X-ray emission.

  16. The Open Flux Problem

    Energy Technology Data Exchange (ETDEWEB)

    Linker, J. A.; Caplan, R. M.; Downs, C.; Riley, P.; Mikic, Z.; Lionello, R. [Predictive Science Inc., 9990 Mesa Rim Road, Suite 170, San Diego, CA 92121 (United States); Henney, C. J. [Air Force Research Lab/Space Vehicles Directorate, 3550 Aberdeen Avenue SE, Kirtland AFB, NM (United States); Arge, C. N. [Science and Exploration Directorate, NASA/GSFC, Greenbelt, MD 20771 (United States); Liu, Y. [W. W. Hansen Experimental Physics Laboratory, Stanford University, Stanford, CA 94305 (United States); Derosa, M. L. [Lockheed Martin Solar and Astrophysics Laboratory, 3251 Hanover Street B/252, Palo Alto, CA 94304 (United States); Yeates, A. [Department of Mathematical Sciences, Durham University, Durham, DH1 3LE (United Kingdom); Owens, M. J., E-mail: linkerj@predsci.com [Space and Atmospheric Electricity Group, Department of Meteorology, University of Reading, Earley Gate, P.O. Box 243, Reading RG6 6BB (United Kingdom)

    2017-10-10

    The heliospheric magnetic field is of pivotal importance in solar and space physics. The field is rooted in the Sun’s photosphere, where it has been observed for many years. Global maps of the solar magnetic field based on full-disk magnetograms are commonly used as boundary conditions for coronal and solar wind models. Two primary observational constraints on the models are (1) the open field regions in the model should approximately correspond to coronal holes (CHs) observed in emission and (2) the magnitude of the open magnetic flux in the model should match that inferred from in situ spacecraft measurements. In this study, we calculate both magnetohydrodynamic and potential field source surface solutions using 14 different magnetic maps produced from five different types of observatory magnetograms, for the time period surrounding 2010 July. We have found that for all of the model/map combinations, models that have CH areas close to observations underestimate the interplanetary magnetic flux, or, conversely, for models to match the interplanetary flux, the modeled open field regions are larger than CHs observed in EUV emission. In an alternative approach, we estimate the open magnetic flux entirely from solar observations by combining automatically detected CHs for Carrington rotation 2098 with observatory synoptic magnetic maps. This approach also underestimates the interplanetary magnetic flux. Our results imply that either typical observatory maps underestimate the Sun’s magnetic flux, or a significant portion of the open magnetic flux is not rooted in regions that are obviously dark in EUV and X-ray emission.

  17. Meromorphic flux compactification

    Energy Technology Data Exchange (ETDEWEB)

    Damian, Cesar [Departamento de Ingeniería Mecánica, Universidad de Guanajuato,Carretera Salamanca-Valle de Santiago Km 3.5+1.8 Comunidad de Palo Blanco,Salamanca (Mexico); Loaiza-Brito, Oscar [Departamento de Física, Universidad de Guanajuato,Loma del Bosque No. 103 Col. Lomas del Campestre C.P 37150 León, Guanajuato (Mexico)

    2017-04-26

    We present exact solutions of four-dimensional Einstein’s equations related to Minkoswki vacuum constructed from Type IIB string theory with non-trivial fluxes. Following https://www.doi.org/10.1007/JHEP02(2015)187; https://www.doi.org/10.1007/JHEP02(2015)188 we study a non-trivial flux compactification on a fibered product by a four-dimensional torus and a two-dimensional sphere punctured by 5- and 7-branes. By considering only 3-form fluxes and the dilaton, as functions on the internal sphere coordinates, we show that these solutions correspond to a family of supersymmetric solutions constructed by the use of G-theory. Meromorphicity on functions constructed in terms of fluxes and warping factors guarantees that flux and 5-brane contributions to the scalar curvature vanish while fulfilling stringent constraints as tadpole cancelation and Bianchi identities. Different Einstein’s solutions are shown to be related by U-dualities. We present three supersymmetric non-trivial Minkowski vacuum solutions and compute the corresponding soft terms. We also construct a non-supersymmetric solution and study its stability.

  18. Meromorphic flux compactification

    International Nuclear Information System (INIS)

    Damian, Cesar; Loaiza-Brito, Oscar

    2017-01-01

    We present exact solutions of four-dimensional Einstein’s equations related to Minkoswki vacuum constructed from Type IIB string theory with non-trivial fluxes. Following https://www.doi.org/10.1007/JHEP02(2015)187; https://www.doi.org/10.1007/JHEP02(2015)188 we study a non-trivial flux compactification on a fibered product by a four-dimensional torus and a two-dimensional sphere punctured by 5- and 7-branes. By considering only 3-form fluxes and the dilaton, as functions on the internal sphere coordinates, we show that these solutions correspond to a family of supersymmetric solutions constructed by the use of G-theory. Meromorphicity on functions constructed in terms of fluxes and warping factors guarantees that flux and 5-brane contributions to the scalar curvature vanish while fulfilling stringent constraints as tadpole cancelation and Bianchi identities. Different Einstein’s solutions are shown to be related by U-dualities. We present three supersymmetric non-trivial Minkowski vacuum solutions and compute the corresponding soft terms. We also construct a non-supersymmetric solution and study its stability.

  19. Flux Pinning in Superconductors

    CERN Document Server

    Matsushita, Teruo

    2007-01-01

    The book covers the flux pinning mechanisms and properties and the electromagnetic phenomena caused by the flux pinning common for metallic, high-Tc and MgB2 superconductors. The condensation energy interaction known for normal precipitates or grain boundaries and the kinetic energy interaction proposed for artificial Nb pins in Nb-Ti, etc., are introduced for the pinning mechanism. Summation theories to derive the critical current density are discussed in detail. Irreversible magnetization and AC loss caused by the flux pinning are also discussed. The loss originally stems from the ohmic dissipation of normal electrons in the normal core driven by the electric field induced by the flux motion. The readers will learn why the resultant loss is of hysteresis type in spite of such mechanism. The influence of the flux pinning on the vortex phase diagram in high Tc superconductors is discussed, and the dependencies of the irreversibility field are also described on other quantities such as anisotropy of supercondu...

  20. Statistical segmentation of multidimensional brain datasets

    Science.gov (United States)

    Desco, Manuel; Gispert, Juan D.; Reig, Santiago; Santos, Andres; Pascau, Javier; Malpica, Norberto; Garcia-Barreno, Pedro

    2001-07-01

    This paper presents an automatic segmentation procedure for MRI neuroimages that overcomes part of the problems involved in multidimensional clustering techniques like partial volume effects (PVE), processing speed and difficulty of incorporating a priori knowledge. The method is a three-stage procedure: 1) Exclusion of background and skull voxels using threshold-based region growing techniques with fully automated seed selection. 2) Expectation Maximization algorithms are used to estimate the probability density function (PDF) of the remaining pixels, which are assumed to be mixtures of gaussians. These pixels can then be classified into cerebrospinal fluid (CSF), white matter and grey matter. Using this procedure, our method takes advantage of using the full covariance matrix (instead of the diagonal) for the joint PDF estimation. On the other hand, logistic discrimination techniques are more robust against violation of multi-gaussian assumptions. 3) A priori knowledge is added using Markov Random Field techniques. The algorithm has been tested with a dataset of 30 brain MRI studies (co-registered T1 and T2 MRI). Our method was compared with clustering techniques and with template-based statistical segmentation, using manual segmentation as a gold-standard. Our results were more robust and closer to the gold-standard.

  1. ASSESSING SMALL SAMPLE WAR-GAMING DATASETS

    Directory of Open Access Journals (Sweden)

    W. J. HURLEY

    2013-10-01

    Full Text Available One of the fundamental problems faced by military planners is the assessment of changes to force structure. An example is whether to replace an existing capability with an enhanced system. This can be done directly with a comparison of measures such as accuracy, lethality, survivability, etc. However this approach does not allow an assessment of the force multiplier effects of the proposed change. To gauge these effects, planners often turn to war-gaming. For many war-gaming experiments, it is expensive, both in terms of time and dollars, to generate a large number of sample observations. This puts a premium on the statistical methodology used to examine these small datasets. In this paper we compare the power of three tests to assess population differences: the Wald-Wolfowitz test, the Mann-Whitney U test, and re-sampling. We employ a series of Monte Carlo simulation experiments. Not unexpectedly, we find that the Mann-Whitney test performs better than the Wald-Wolfowitz test. Resampling is judged to perform slightly better than the Mann-Whitney test.

  2. Neutron flux monitoring device

    International Nuclear Information System (INIS)

    Goto, Yasushi; Mitsubori, Minehisa; Ohashi, Kazunori.

    1997-01-01

    The present invention provides a neutron flux monitoring device for preventing occurrence of erroneous reactor scram caused by the elevation of the indication of a start region monitor (SRM) due to a factor different from actual increase of neutron fluxes. Namely, judgement based on measured values obtained by a pulse counting method and a judgment based on measured values obtained by a Cambel method are combined. A logic of switching neutron flux measuring method to be used for monitoring, namely, switching to an intermediate region when both of the judgements are valid is adopted. Then, even if the indication value is elevated based on the Cambel method with no increase of the counter rate in a neutron source region, the switching to the intermediate region is not conducted. As a result, erroneous reactor scram such as 'shorter reactor period' can be avoided. (I.S.)

  3. Atmospheric neutrino fluxes

    International Nuclear Information System (INIS)

    Honda, M.; Kasahara, K.; Hidaka, K.; Midorikawa, S.

    1990-02-01

    A detailed Monte Carlo simulation of neutrino fluxes of atmospheric origin is made taking into account the muon polarization effect on neutrinos from muon decay. We calculate the fluxes with energies above 3 MeV for future experiments. There still remains a significant discrepancy between the calculated (ν e +antiν e )/(ν μ +antiν μ ) ratio and that observed by the Kamiokande group. However, the ratio evaluated at the Frejus site shows a good agreement with the data. (author)

  4. Climate Model Evaluation using New Datasets from the Clouds and the Earth's Radiant Energy System (CERES)

    Science.gov (United States)

    Loeb, Norman G.; Wielicki, Bruce A.; Doelling, David R.

    2008-01-01

    There are some in the science community who believe that the response of the climate system to anthropogenic radiative forcing is unpredictable and we should therefore call off the quest . The key limitation in climate predictability is associated with cloud feedback. Narrowing the uncertainty in cloud feedback (and therefore climate sensitivity) requires optimal use of the best available observations to evaluate and improve climate model processes and constrain climate model simulations over longer time scales. The Clouds and the Earth s Radiant Energy System (CERES) is a satellite-based program that provides global cloud, aerosol and radiative flux observations for improving our understanding of cloud-aerosol-radiation feedbacks in the Earth s climate system. CERES is the successor to the Earth Radiation Budget Experiment (ERBE), which has widely been used to evaluate climate models both at short time scales (e.g., process studies) and at decadal time scales. A CERES instrument flew on the TRMM satellite and captured the dramatic 1998 El Nino, and four other CERES instruments are currently flying aboard the Terra and Aqua platforms. Plans are underway to fly the remaining copy of CERES on the upcoming NPP spacecraft (mid-2010 launch date). Every aspect of CERES represents a significant improvement over ERBE. While both CERES and ERBE measure broadband radiation, CERES calibration is a factor of 2 better than ERBE. In order to improve the characterization of clouds and aerosols within a CERES footprint, we use coincident higher-resolution imager observations (VIRS, MODIS or VIIRS) to provide a consistent cloud-aerosol-radiation dataset at climate accuracy. Improved radiative fluxes are obtained by using new CERES-derived Angular Distribution Models (ADMs) for converting measured radiances to fluxes. CERES radiative fluxes are a factor of 2 more accurate than ERBE overall, but the improvement by cloud type and at high latitudes can be as high as a factor of 5

  5. Radiation flux measuring device

    International Nuclear Information System (INIS)

    Corte, E.; Maitra, P.

    1977-01-01

    A radiation flux measuring device is described which employs a differential pair of transistors, the output of which is maintained constant, connected to a radiation detector. Means connected to the differential pair produce a signal representing the log of the a-c component of the radiation detector, thereby providing a signal representing the true root mean square logarithmic output. 3 claims, 2 figures

  6. Soluble organic nutrient fluxes

    Science.gov (United States)

    Robert G. Qualls; Bruce L. Haines; Wayne Swank

    2014-01-01

    Our objectives in this study were (i) compare fluxes of the dissolved organic nutrients dissolved organic carbon (DOC), DON, and dissolved organic phosphorus (DOP) in a clearcut area and an adjacent mature reference area. (ii) determine whether concentrations of dissolved organic nutrients or inorganic nutrients were greater in clearcut areas than in reference areas,...

  7. Flux vacua and supermanifolds

    Energy Technology Data Exchange (ETDEWEB)

    Grassi, Pietro Antonio [CERN, Theory Unit, CH-1211 Geneva, 23 (Switzerland); Marescotti, Matteo [Dipartimento di Fisica Teorica, Universita di Torino, Via Giuria 1, I-10125, Turin (Italy)

    2007-01-15

    As been recently pointed out, physically relevant models derived from string theory require the presence of non-vanishing form fluxes besides the usual geometrical constraints. In the case of NS-NS fluxes, the Generalized Complex Geometry encodes these informations in a beautiful geometrical structure. On the other hand, the R-R fluxes call for supergeometry as the underlying mathematical framework. In this context, we analyze the possibility of constructing interesting supermanifolds recasting the geometrical data and RR fluxes. To characterize these supermanifolds we have been guided by the fact topological strings on supermanifolds require the super-Ricci flatness of the target space. This can be achieved by adding to a given bosonic manifold enough anticommuting coordinates and new constraints on the bosonic sub-manifold. We study these constraints at the linear and non-linear level for a pure geometrical setting and in the presence of p-form field strengths. We find that certain spaces admit several super-extensions and we give a parameterization in a simple case of d bosonic coordinates and two fermionic coordinates. In addition, we comment on the role of the RR field in the construction of the super-metric. We give several examples based on supergroup manifolds and coset supermanifolds.

  8. Flux vacua and supermanifolds

    International Nuclear Information System (INIS)

    Grassi, Pietro Antonio; Marescotti, Matteo

    2007-01-01

    As been recently pointed out, physically relevant models derived from string theory require the presence of non-vanishing form fluxes besides the usual geometrical constraints. In the case of NS-NS fluxes, the Generalized Complex Geometry encodes these informations in a beautiful geometrical structure. On the other hand, the R-R fluxes call for supergeometry as the underlying mathematical framework. In this context, we analyze the possibility of constructing interesting supermanifolds recasting the geometrical data and RR fluxes. To characterize these supermanifolds we have been guided by the fact topological strings on supermanifolds require the super-Ricci flatness of the target space. This can be achieved by adding to a given bosonic manifold enough anticommuting coordinates and new constraints on the bosonic sub-manifold. We study these constraints at the linear and non-linear level for a pure geometrical setting and in the presence of p-form field strengths. We find that certain spaces admit several super-extensions and we give a parameterization in a simple case of d bosonic coordinates and two fermionic coordinates. In addition, we comment on the role of the RR field in the construction of the super-metric. We give several examples based on supergroup manifolds and coset supermanifolds

  9. Atmospheric neutrino fluxes

    International Nuclear Information System (INIS)

    Perkins, D.H.

    1984-01-01

    The atmospheric neutrino fluxes, which are responsible for the main background in proton decay experiments, have been calculated by two independent methods. There are discrepancies between the two sets of results regarding latitude effects and up-down asymmetries, especially for neutrino energies Esub(ν) < 1 GeV. (author)

  10. Flux scaling: Ultimate regime

    Indian Academy of Sciences (India)

    First page Back Continue Last page Overview Graphics. Flux scaling: Ultimate regime. With the Nusselt number and the mixing length scales, we get the Nusselt number and Reynolds number (w'd/ν) scalings: and or. and. scaling expected to occur at extremely high Ra Rayleigh-Benard convection. Get the ultimate regime ...

  11. The Dataset of Countries at Risk of Electoral Violence

    OpenAIRE

    Birch, Sarah; Muchlinski, David

    2017-01-01

    Electoral violence is increasingly affecting elections around the world, yet researchers have been limited by a paucity of granular data on this phenomenon. This paper introduces and describes a new dataset of electoral violence – the Dataset of Countries at Risk of Electoral Violence (CREV) – that provides measures of 10 different types of electoral violence across 642 elections held around the globe between 1995 and 2013. The paper provides a detailed account of how and why the dataset was ...

  12. Norwegian Hydrological Reference Dataset for Climate Change Studies

    Energy Technology Data Exchange (ETDEWEB)

    Magnussen, Inger Helene; Killingland, Magnus; Spilde, Dag

    2012-07-01

    Based on the Norwegian hydrological measurement network, NVE has selected a Hydrological Reference Dataset for studies of hydrological change. The dataset meets international standards with high data quality. It is suitable for monitoring and studying the effects of climate change on the hydrosphere and cryosphere in Norway. The dataset includes streamflow, groundwater, snow, glacier mass balance and length change, lake ice and water temperature in rivers and lakes.(Author)

  13. Public Availability to ECS Collected Datasets

    Science.gov (United States)

    Henderson, J. F.; Warnken, R.; McLean, S. J.; Lim, E.; Varner, J. D.

    2013-12-01

    Coastal nations have spent considerable resources exploring the limits of their extended continental shelf (ECS) beyond 200 nm. Although these studies are funded to fulfill requirements of the UN Convention on the Law of the Sea, the investments are producing new data sets in frontier areas of Earth's oceans that will be used to understand, explore, and manage the seafloor and sub-seafloor for decades to come. Although many of these datasets are considered proprietary until a nation's potential ECS has become 'final and binding' an increasing amount of data are being released and utilized by the public. Data sets include multibeam, seismic reflection/refraction, bottom sampling, and geophysical data. The U.S. ECS Project, a multi-agency collaboration whose mission is to establish the full extent of the continental shelf of the United States consistent with international law, relies heavily on data and accurate, standard metadata. The United States has made it a priority to make available to the public all data collected with ECS-funding as quickly as possible. The National Oceanic and Atmospheric Administration's (NOAA) National Geophysical Data Center (NGDC) supports this objective by partnering with academia and other federal government mapping agencies to archive, inventory, and deliver marine mapping data in a coordinated, consistent manner. This includes ensuring quality, standard metadata and developing and maintaining data delivery capabilities built on modern digital data archives. Other countries, such as Ireland, have submitted their ECS data for public availability and many others have made pledges to participate in the future. The data services provided by NGDC support the U.S. ECS effort as well as many developing nation's ECS effort through the U.N. Environmental Program. Modern discovery, visualization, and delivery of scientific data and derived products that span national and international sources of data ensure the greatest re-use of data and

  14. BIA Indian Lands Dataset (Indian Lands of the United States)

    Data.gov (United States)

    Federal Geographic Data Committee — The American Indian Reservations / Federally Recognized Tribal Entities dataset depicts feature location, selected demographics and other associated data for the 561...

  15. Framework for Interactive Parallel Dataset Analysis on the Grid

    Energy Technology Data Exchange (ETDEWEB)

    Alexander, David A.; Ananthan, Balamurali; /Tech-X Corp.; Johnson, Tony; Serbo, Victor; /SLAC

    2007-01-10

    We present a framework for use at a typical Grid site to facilitate custom interactive parallel dataset analysis targeting terabyte-scale datasets of the type typically produced by large multi-institutional science experiments. We summarize the needs for interactive analysis and show a prototype solution that satisfies those needs. The solution consists of desktop client tool and a set of Web Services that allow scientists to sign onto a Grid site, compose analysis script code to carry out physics analysis on datasets, distribute the code and datasets to worker nodes, collect the results back to the client, and to construct professional-quality visualizations of the results.

  16. Socioeconomic Data and Applications Center (SEDAC) Treaty Status Dataset

    Data.gov (United States)

    National Aeronautics and Space Administration — The Socioeconomic Data and Application Center (SEDAC) Treaty Status Dataset contains comprehensive treaty information for multilateral environmental agreements,...

  17. Design of a flux buffer based on the flux shuttle

    International Nuclear Information System (INIS)

    Gershenson, M.

    1991-01-01

    This paper discusses the design considerations for a flux buffer based on the flux-shuttle concept. Particular attention is given to the issues of flux popping, stability of operation and saturation levels for a large input. Modulation techniques used in order to minimize 1/f noise, in addition to offsets are also analyzed. Advantages over conventional approaches using a SQUID for a flux buffer are discussed. Results of computer simulations are presented

  18. Lunar Meteorites: A Global Geochemical Dataset

    Science.gov (United States)

    Zeigler, R. A.; Joy, K. H.; Arai, T.; Gross, J.; Korotev, R. L.; McCubbin, F. M.

    2017-01-01

    To date, the world's meteorite collections contain over 260 lunar meteorite stones representing at least 120 different lunar meteorites. Additionally, there are 20-30 as yet unnamed stones currently in the process of being classified. Collectively these lunar meteorites likely represent 40-50 distinct sampling locations from random locations on the Moon. Although the exact provenance of each individual lunar meteorite is unknown, collectively the lunar meteorites represent the best global average of the lunar crust. The Apollo sites are all within or near the Procellarum KREEP Terrane (PKT), thus lithologies from the PKT are overrepresented in the Apollo sample suite. Nearly all of the lithologies present in the Apollo sample suite are found within the lunar meteorites (high-Ti basalts are a notable exception), and the lunar meteorites contain several lithologies not present in the Apollo sample suite (e.g., magnesian anorthosite). This chapter will not be a sample-by-sample summary of each individual lunar meteorite. Rather, the chapter will summarize the different types of lunar meteorites and their relative abundances, comparing and contrasting the lunar meteorite sample suite with the Apollo sample suite. This chapter will act as one of the introductory chapters to the volume, introducing lunar samples in general and setting the stage for more detailed discussions in later more specialized chapters. The chapter will begin with a description of how lunar meteorites are ejected from the Moon, how deep samples are being excavated from, what the likely pairing relationships are among the lunar meteorite samples, and how the lunar meteorites can help to constrain the impactor flux in the inner solar system. There will be a discussion of the biases inherent to the lunar meteorite sample suite in terms of underrepresented lithologies or regions of the Moon, and an examination of the contamination and limitations of lunar meteorites due to terrestrial weathering. The

  19. Lobotomy of flux compactifications

    Energy Technology Data Exchange (ETDEWEB)

    Dibitetto, Giuseppe [Institutionen för fysik och astronomi, University of Uppsala,Box 803, SE-751 08 Uppsala (Sweden); Guarino, Adolfo [Albert Einstein Center for Fundamental Physics, Institute for Theoretical Physics,Bern University, Sidlerstrasse 5, CH-3012 Bern (Switzerland); Roest, Diederik [Centre for Theoretical Physics, University of Groningen,Nijenborgh 4 9747 AG Groningen (Netherlands)

    2014-05-15

    We provide the dictionary between four-dimensional gauged supergravity and type II compactifications on T{sup 6} with metric and gauge fluxes in the absence of supersymmetry breaking sources, such as branes and orientifold planes. Secondly, we prove that there is a unique isotropic compactification allowing for critical points. It corresponds to a type IIA background given by a product of two 3-tori with SO(3) twists and results in a unique theory (gauging) with a non-semisimple gauge algebra. Besides the known four AdS solutions surviving the orientifold projection to N=4 induced by O6-planes, this theory contains a novel AdS solution that requires non-trivial orientifold-odd fluxes, hence being a genuine critical point of the N=8 theory.

  20. Derivation of ozone flux-yield relationships for lettuce: A key horticultural crop

    International Nuclear Information System (INIS)

    Goumenaki, Eleni; Fernandez, Ignacio Gonzalez; Papanikolaou, Antigoni; Papadopoulou, Despoina; Askianakis, Christos; Kouvarakis, George; Barnes, Jeremy

    2007-01-01

    Ozone flux-response relationships were derived for lettuce, employing a multiplicative approach to model the manner in which stomatal conductance is influenced by key environmental variables, using a dataset collected during field experimentation in Crete and yield-response relationships derived from parallel open-top chamber experiments. Regional agronomic practices were adopted throughout. Computed versus measured data revealed that the derived model explained 51% (P -2 s -1 . Regressions employing very low or zero flux thresholds resulted in the strongest yield-flux relationships (explaining ∼80% (P < 0.05) of the variation in the dataset). - Establishment of ozone flux-yield relationships for a commercially-important horticultural crop grown widely in the Mediterranean

  1. Physics of magnetic flux ropes

    Science.gov (United States)

    Russell, C. T.; Priest, E. R.; Lee, L. C.

    The present work encompasses papers on the structure, waves, and instabilities of magnetic flux ropes (MFRs), photospheric flux tubes (PFTs), the structure and heating of coronal loops, solar prominences, coronal mass ejections and magnetic clouds, flux ropes in planetary ionospheres, the magnetopause, magnetospheric field-aligned currents and flux tubes, and the magnetotail. Attention is given to the equilibrium of MFRs, resistive instability, magnetic reconnection and turbulence in current sheets, dynamical effects and energy transport in intense flux tubes, waves in solar PFTs, twisted flux ropes in the solar corona, an electrodynamical model of solar flares, filament cooling and condensation in a sheared magnetic field, the magnetopause, the generation of twisted MFRs during magnetic reconnection, ionospheric flux ropes above the South Pole, substorms and MFR structures, evidence for flux ropes in the earth magnetotail, and MFRs in 3D MHD simulations.

  2. An Analysis of the GTZAN Music Genre Dataset

    DEFF Research Database (Denmark)

    Sturm, Bob L.

    2012-01-01

    Most research in automatic music genre recognition has used the dataset assembled by Tzanetakis et al. in 2001. The composition and integrity of this dataset, however, has never been formally analyzed. For the first time, we provide an analysis of its composition, and create a machine...

  3. Really big data: Processing and analysis of large datasets

    Science.gov (United States)

    Modern animal breeding datasets are large and getting larger, due in part to the recent availability of DNA data for many animals. Computational methods for efficiently storing and analyzing those data are under development. The amount of storage space required for such datasets is increasing rapidl...

  4. An Annotated Dataset of 14 Cardiac MR Images

    DEFF Research Database (Denmark)

    Stegmann, Mikkel Bille

    2002-01-01

    This note describes a dataset consisting of 14 annotated cardiac MR images. Points of correspondence are placed on each image at the left ventricle (LV). As such, the dataset can be readily used for building statistical models of shape. Further, format specifications and terms of use are given....

  5. A New Outlier Detection Method for Multidimensional Datasets

    KAUST Repository

    Abdel Messih, Mario A.

    2012-07-01

    This study develops a novel hybrid method for outlier detection (HMOD) that combines the idea of distance based and density based methods. The proposed method has two main advantages over most of the other outlier detection methods. The first advantage is that it works well on both dense and sparse datasets. The second advantage is that, unlike most other outlier detection methods that require careful parameter setting and prior knowledge of the data, HMOD is not very sensitive to small changes in parameter values within certain parameter ranges. The only required parameter to set is the number of nearest neighbors. In addition, we made a fully parallelized implementation of HMOD that made it very efficient in applications. Moreover, we proposed a new way of using the outlier detection for redundancy reduction in datasets where the confidence level that evaluates how accurate the less redundant dataset can be used to represent the original dataset can be specified by users. HMOD is evaluated on synthetic datasets (dense and mixed “dense and sparse”) and a bioinformatics problem of redundancy reduction of dataset of position weight matrices (PWMs) of transcription factor binding sites. In addition, in the process of assessing the performance of our redundancy reduction method, we developed a simple tool that can be used to evaluate the confidence level of reduced dataset representing the original dataset. The evaluation of the results shows that our method can be used in a wide range of problems.

  6. ATLAS File and Dataset Metadata Collection and Use

    CERN Document Server

    Albrand, S; The ATLAS collaboration; Lambert, F; Gallas, E J

    2012-01-01

    The ATLAS Metadata Interface (“AMI”) was designed as a generic cataloguing system, and as such it has found many uses in the experiment including software release management, tracking of reconstructed event sizes and control of dataset nomenclature. The primary use of AMI is to provide a catalogue of datasets (file collections) which is searchable using physics criteria. In this paper we discuss the various mechanisms used for filling the AMI dataset and file catalogues. By correlating information from different sources we can derive aggregate information which is important for physics analysis; for example the total number of events contained in dataset, and possible reasons for missing events such as a lost file. Finally we will describe some specialized interfaces which were developed for the Data Preparation and reprocessing coordinators. These interfaces manipulate information from both the dataset domain held in AMI, and the run-indexed information held in the ATLAS COMA application (Conditions and ...

  7. A dataset on tail risk of commodities markets.

    Science.gov (United States)

    Powell, Robert J; Vo, Duc H; Pham, Thach N; Singh, Abhay K

    2017-12-01

    This article contains the datasets related to the research article "The long and short of commodity tails and their relationship to Asian equity markets"(Powell et al., 2017) [1]. The datasets contain the daily prices (and price movements) of 24 different commodities decomposed from the S&P GSCI index and the daily prices (and price movements) of three share market indices including World, Asia, and South East Asia for the period 2004-2015. Then, the dataset is divided into annual periods, showing the worst 5% of price movements for each year. The datasets are convenient to examine the tail risk of different commodities as measured by Conditional Value at Risk (CVaR) as well as their changes over periods. The datasets can also be used to investigate the association between commodity markets and share markets.

  8. Surface Turbulent Fluxes, 1x1 deg Monthly Grid, Set1 and Interpolated Data V2c (GSSTFM) at GES DISC

    Data.gov (United States)

    National Aeronautics and Space Administration — These data are the Goddard Satellite-based Surface Turbulent Fluxes Version-2c Dataset recently produced through a MEaSUREs funded project led by Dr. Chung-Lin Shie...

  9. Australian methane fluxes

    International Nuclear Information System (INIS)

    Williams, D.J.

    1990-01-01

    Estimates are provided for the amount of methane emitted annually into the atmosphere in Australia for a variety of sources. The sources considered are coal mining, landfill, motor vehicles, natural gas suply system, rice paddies, bushfires, termites, wetland and animals. This assessment indicates that the major sources of methane are natural or agricultural in nature and therefore offer little scope for reduction. Nevertheless the remainder are not trival and reduction of these fluxes could play a significant part in any Australian action on the greenhouse problem. 19 refs., 7 tabs., 1 fig

  10. Discovery and Reuse of Open Datasets: An Exploratory Study

    Directory of Open Access Journals (Sweden)

    Sara

    2016-07-01

    Full Text Available Objective: This article analyzes twenty cited or downloaded datasets and the repositories that house them, in order to produce insights that can be used by academic libraries to encourage discovery and reuse of research data in institutional repositories. Methods: Using Thomson Reuters’ Data Citation Index and repository download statistics, we identified twenty cited/downloaded datasets. We documented the characteristics of the cited/downloaded datasets and their corresponding repositories in a self-designed rubric. The rubric includes six major categories: basic information; funding agency and journal information; linking and sharing; factors to encourage reuse; repository characteristics; and data description. Results: Our small-scale study suggests that cited/downloaded datasets generally comply with basic recommendations for facilitating reuse: data are documented well; formatted for use with a variety of software; and shared in established, open access repositories. Three significant factors also appear to contribute to dataset discovery: publishing in discipline-specific repositories; indexing in more than one location on the web; and using persistent identifiers. The cited/downloaded datasets in our analysis came from a few specific disciplines, and tended to be funded by agencies with data publication mandates. Conclusions: The results of this exploratory research provide insights that can inform academic librarians as they work to encourage discovery and reuse of institutional datasets. Our analysis also suggests areas in which academic librarians can target open data advocacy in their communities in order to begin to build open data success stories that will fuel future advocacy efforts.

  11. Viability of Controlling Prosthetic Hand Utilizing Electroencephalograph (EEG) Dataset Signal

    Science.gov (United States)

    Miskon, Azizi; A/L Thanakodi, Suresh; Raihan Mazlan, Mohd; Mohd Haziq Azhar, Satria; Nooraya Mohd Tawil, Siti

    2016-11-01

    This project presents the development of an artificial hand controlled by Electroencephalograph (EEG) signal datasets for the prosthetic application. The EEG signal datasets were used as to improvise the way to control the prosthetic hand compared to the Electromyograph (EMG). The EMG has disadvantages to a person, who has not used the muscle for a long time and also to person with degenerative issues due to age factor. Thus, the EEG datasets found to be an alternative for EMG. The datasets used in this work were taken from Brain Computer Interface (BCI) Project. The datasets were already classified for open, close and combined movement operations. It served the purpose as an input to control the prosthetic hand by using an Interface system between Microsoft Visual Studio and Arduino. The obtained results reveal the prosthetic hand to be more efficient and faster in response to the EEG datasets with an additional LiPo (Lithium Polymer) battery attached to the prosthetic. Some limitations were also identified in terms of the hand movements, weight of the prosthetic, and the suggestions to improve were concluded in this paper. Overall, the objective of this paper were achieved when the prosthetic hand found to be feasible in operation utilizing the EEG datasets.

  12. Sparse Group Penalized Integrative Analysis of Multiple Cancer Prognosis Datasets

    Science.gov (United States)

    Liu, Jin; Huang, Jian; Xie, Yang; Ma, Shuangge

    2014-01-01

    SUMMARY In cancer research, high-throughput profiling studies have been extensively conducted, searching for markers associated with prognosis. Because of the “large d, small n” characteristic, results generated from the analysis of a single dataset can be unsatisfactory. Recent studies have shown that integrative analysis, which simultaneously analyzes multiple datasets, can be more effective than single-dataset analysis and classic meta-analysis. In most of existing integrative analysis, the homogeneity model has been assumed, which postulates that different datasets share the same set of markers. Several approaches have been designed to reinforce this assumption. In practice, different datasets may differ in terms of patient selection criteria, profiling techniques, and many other aspects. Such differences may make the homogeneity model too restricted. In this study, we assume the heterogeneity model, under which different datasets are allowed to have different sets of markers. With multiple cancer prognosis datasets, we adopt the AFT (accelerated failure time) model to describe survival. This model may have the lowest computational cost among popular semiparametric survival models. For marker selection, we adopt a sparse group MCP (minimax concave penalty) approach. This approach has an intuitive formulation and can be computed using an effective group coordinate descent algorithm. Simulation study shows that it outperforms the existing approaches under both the homogeneity and heterogeneity models. Data analysis further demonstrates the merit of heterogeneity model and proposed approach. PMID:23938111

  13. PROVIDING GEOGRAPHIC DATASETS AS LINKED DATA IN SDI

    Directory of Open Access Journals (Sweden)

    E. Hietanen

    2016-06-01

    Full Text Available In this study, a prototype service to provide data from Web Feature Service (WFS as linked data is implemented. At first, persistent and unique Uniform Resource Identifiers (URI are created to all spatial objects in the dataset. The objects are available from those URIs in Resource Description Framework (RDF data format. Next, a Web Ontology Language (OWL ontology is created to describe the dataset information content using the Open Geospatial Consortium’s (OGC GeoSPARQL vocabulary. The existing data model is modified in order to take into account the linked data principles. The implemented service produces an HTTP response dynamically. The data for the response is first fetched from existing WFS. Then the Geographic Markup Language (GML format output of the WFS is transformed on-the-fly to the RDF format. Content Negotiation is used to serve the data in different RDF serialization formats. This solution facilitates the use of a dataset in different applications without replicating the whole dataset. In addition, individual spatial objects in the dataset can be referred with URIs. Furthermore, the needed information content of the objects can be easily extracted from the RDF serializations available from those URIs. A solution for linking data objects to the dataset URI is also introduced by using the Vocabulary of Interlinked Datasets (VoID. The dataset is divided to the subsets and each subset is given its persistent and unique URI. This enables the whole dataset to be explored with a web browser and all individual objects to be indexed by search engines.

  14. Homogenised Australian climate datasets used for climate change monitoring

    International Nuclear Information System (INIS)

    Trewin, Blair; Jones, David; Collins; Dean; Jovanovic, Branislava; Braganza, Karl

    2007-01-01

    Full text: The Australian Bureau of Meteorology has developed a number of datasets for use in climate change monitoring. These datasets typically cover 50-200 stations distributed as evenly as possible over the Australian continent, and have been subject to detailed quality control and homogenisation.The time period over which data are available for each element is largely determined by the availability of data in digital form. Whilst nearly all Australian monthly and daily precipitation data have been digitised, a significant quantity of pre-1957 data (for temperature and evaporation) or pre-1987 data (for some other elements) remains to be digitised, and is not currently available for use in the climate change monitoring datasets. In the case of temperature and evaporation, the start date of the datasets is also determined by major changes in instruments or observing practices for which no adjustment is feasible at the present time. The datasets currently available cover: Monthly and daily precipitation (most stations commence 1915 or earlier, with many extending back to the late 19th century, and a few to the mid-19th century); Annual temperature (commences 1910); Daily temperature (commences 1910, with limited station coverage pre-1957); Twice-daily dewpoint/relative humidity (commences 1957); Monthly pan evaporation (commences 1970); Cloud amount (commences 1957) (Jovanovic etal. 2007). As well as the station-based datasets listed above, an additional dataset being developed for use in climate change monitoring (and other applications) covers tropical cyclones in the Australian region. This is described in more detail in Trewin (2007). The datasets already developed are used in analyses of observed climate change, which are available through the Australian Bureau of Meteorology website (http://www.bom.gov.au/silo/products/cli_chg/). They are also used as a basis for routine climate monitoring, and in the datasets used for the development of seasonal

  15. AmeriFlux Site and Data Exploration System

    Science.gov (United States)

    Krassovski, M.; Boden, T.; Yang, B.; Jackson, B.

    2011-12-01

    The AmeriFlux network was established in 1996. The network provides continuous observations of ecosystem-level exchanges of CO2, water, energy and momentum spanning diurnal, synoptic, seasonal, and interannual time scales. The current network, including both active and inactive sites, consists of 141 sites in North, Central, and South America. The Carbon Dioxide Information Analysis Center (CDIAC) at Oak Ridge National Laboratory (ORNL) provides data management support for the AmeriFlux network including long-term data storage and dissemination. AmeriFlux offers a broad suite of value-added data products: Level 1 data products at 30 minute or hourly time intervals provided by the site teams, Level 2 data processed by CDIAC and Level 3 and 4 files created using CarboEurope algorithms. CDIAC has developed a relational database to house the vast array of AmeriFlux data and information and a web-based interface to the database, the AmeriFlux Site and Data Exploration System (http://ameriflux.ornl.gov), to help users worldwide identify, and more recently, download desired AmeriFlux data. AmeriFlux and CDIAC offer numerous value-added AmeriFlux data products (i.e., Level 1-4 data products, biological data) and most of these data products are or will be available through the new data system. Vital site information (e.g., location coordinates, dominant species, land-use history) is also displayed in the new system. The data system provides numerous ways to explore and extract data. Searches can be done by site, location, measurement status, available data products, vegetation types, and by reported measurements just to name a few. Data can be accessed through the links to full data sets reported by a site, organized by types of data products, or by creating customized datasets based on user search criteria. The new AmeriFlux download module contains features intended to ease compliance of the AmeriFlux fair-use data policy, acknowledge the contributions of submitting

  16. Tension in the recent Type Ia supernovae datasets

    International Nuclear Information System (INIS)

    Wei, Hao

    2010-01-01

    In the present work, we investigate the tension in the recent Type Ia supernovae (SNIa) datasets Constitution and Union. We show that they are in tension not only with the observations of the cosmic microwave background (CMB) anisotropy and the baryon acoustic oscillations (BAO), but also with other SNIa datasets such as Davis and SNLS. Then, we find the main sources responsible for the tension. Further, we make this more robust by employing the method of random truncation. Based on the results of this work, we suggest two truncated versions of the Union and Constitution datasets, namely the UnionT and ConstitutionT SNIa samples, whose behaviors are more regular.

  17. Critical heat flux evaluation

    International Nuclear Information System (INIS)

    Banner, D.

    1995-01-01

    Critical heat flux (CHF) is of importance for nuclear safety and represents the major limiting factors for reactor cores. Critical heat flux is caused by a sharp reduction in the heat transfer coefficient located at the outer surface of fuel rods. Safety requires that this phenomenon also called the boiling crisis should be precluded under nominal or incidental conditions (Class I and II events). CHF evaluation in reactor cores is basically a two-step approach. Fuel assemblies are first tested in experimental loops in order to determine CHF limits under various flow conditions. Then, core thermal-hydraulic calculations are performed for safety evaluation. The paper will go into more details about the boiling crisis in order to pinpoint complexity and lack of fundamental understanding in many areas. Experimental test sections needed to collect data over wide thermal-hydraulic and geometric ranges are described CHF safety margin evaluation in reactors cores is discussed by presenting how uncertainties are mentioned. From basic considerations to current concerns, the following topics are discussed; knowledge of the boiling crisis, CHF predictors, and advances thermal-hydraulic codes. (authors). 15 refs., 4 figs

  18. Neutron flux monitor

    International Nuclear Information System (INIS)

    Seki, Eiji; Tai, Ichiro.

    1984-01-01

    Purpose: To maintain the measuring accuracy and the reponse time within an allowable range in accordance with the change of neutron fluxes in a nuclear reactor pressure vessel. Constitution: Neutron fluxes within a nuclear reactor pressure vessel are detected by detectors, converted into pulse signals and amplified in a range switching amplifier. The amplified signals are further converted through an A/D converter and digital signals from the converter are subjected to a square operation in an square operation circuit. The output from the circuit is inputted into an integration circuit to selectively accumulate the constant of 1/2n, 1 - 1/2n (n is a positive integer) respectively for two continuing signals to perform weighing. Then, the addition is carried out to calculate the integrated value and the addition number is changed by the chane in the number n to vary the integrating time. The integrated value is inputted into a control circuit to control the value of n so that the fluctuation and the calculation time for the integrated value are within a predetermined range and, at the same time, the gain of the range switching amplifier is controlled. (Seki, T.)

  19. Background qualitative analysis of the European reference life cycle database (ELCD) energy datasets - part II: electricity datasets.

    Science.gov (United States)

    Garraín, Daniel; Fazio, Simone; de la Rúa, Cristina; Recchioni, Marco; Lechón, Yolanda; Mathieux, Fabrice

    2015-01-01

    The aim of this paper is to identify areas of potential improvement of the European Reference Life Cycle Database (ELCD) electricity datasets. The revision is based on the data quality indicators described by the International Life Cycle Data system (ILCD) Handbook, applied on sectorial basis. These indicators evaluate the technological, geographical and time-related representativeness of the dataset and the appropriateness in terms of completeness, precision and methodology. Results show that ELCD electricity datasets have a very good quality in general terms, nevertheless some findings and recommendations in order to improve the quality of Life-Cycle Inventories have been derived. Moreover, these results ensure the quality of the electricity-related datasets to any LCA practitioner, and provide insights related to the limitations and assumptions underlying in the datasets modelling. Giving this information, the LCA practitioner will be able to decide whether the use of the ELCD electricity datasets is appropriate based on the goal and scope of the analysis to be conducted. The methodological approach would be also useful for dataset developers and reviewers, in order to improve the overall Data Quality Requirements of databases.

  20. Dataset definition for CMS operations and physics analyses

    Science.gov (United States)

    Franzoni, Giovanni; Compact Muon Solenoid Collaboration

    2016-04-01

    Data recorded at the CMS experiment are funnelled into streams, integrated in the HLT menu, and further organised in a hierarchical structure of primary datasets and secondary datasets/dedicated skims. Datasets are defined according to the final-state particles reconstructed by the high level trigger, the data format and the use case (physics analysis, alignment and calibration, performance studies). During the first LHC run, new workflows have been added to this canonical scheme, to exploit at best the flexibility of the CMS trigger and data acquisition systems. The concepts of data parking and data scouting have been introduced to extend the physics reach of CMS, offering the opportunity of defining physics triggers with extremely loose selections (e.g. dijet resonance trigger collecting data at a 1 kHz). In this presentation, we review the evolution of the dataset definition during the LHC run I, and we discuss the plans for the run II.

  1. U.S. Climate Divisional Dataset (Version Superseded)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This data has been superseded by a newer version of the dataset. Please refer to NOAA's Climate Divisional Database for more information. The U.S. Climate Divisional...

  2. Karna Particle Size Dataset for Tables and Figures

    Data.gov (United States)

    U.S. Environmental Protection Agency — This dataset contains 1) table of bulk Pb-XAS LCF results, 2) table of bulk As-XAS LCF results, 3) figure data of particle size distribution, and 4) figure data for...

  3. NOAA Global Surface Temperature Dataset, Version 4.0

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The NOAA Global Surface Temperature Dataset (NOAAGlobalTemp) is derived from two independent analyses: the Extended Reconstructed Sea Surface Temperature (ERSST)...

  4. National Hydrography Dataset (NHD) - USGS National Map Downloadable Data Collection

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — The USGS National Hydrography Dataset (NHD) Downloadable Data Collection from The National Map (TNM) is a comprehensive set of digital spatial data that encodes...

  5. Watershed Boundary Dataset (WBD) - USGS National Map Downloadable Data Collection

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — The Watershed Boundary Dataset (WBD) from The National Map (TNM) defines the perimeter of drainage areas formed by the terrain and other landscape characteristics....

  6. BASE MAP DATASET, LE FLORE COUNTY, OKLAHOMA, USA

    Data.gov (United States)

    Federal Emergency Management Agency, Department of Homeland Security — Basemap datasets comprise six of the seven FGDC themes of geospatial data that are used by most GIS applications (Note: the seventh framework theme, orthographic...

  7. USGS National Hydrography Dataset from The National Map

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — USGS The National Map - National Hydrography Dataset (NHD) is a comprehensive set of digital spatial data that encodes information about naturally occurring and...

  8. A robust dataset-agnostic heart disease classifier from Phonocardiogram.

    Science.gov (United States)

    Banerjee, Rohan; Dutta Choudhury, Anirban; Deshpande, Parijat; Bhattacharya, Sakyajit; Pal, Arpan; Mandana, K M

    2017-07-01

    Automatic classification of normal and abnormal heart sounds is a popular area of research. However, building a robust algorithm unaffected by signal quality and patient demography is a challenge. In this paper we have analysed a wide list of Phonocardiogram (PCG) features in time and frequency domain along with morphological and statistical features to construct a robust and discriminative feature set for dataset-agnostic classification of normal and cardiac patients. The large and open access database, made available in Physionet 2016 challenge was used for feature selection, internal validation and creation of training models. A second dataset of 41 PCG segments, collected using our in-house smart phone based digital stethoscope from an Indian hospital was used for performance evaluation. Our proposed methodology yielded sensitivity and specificity scores of 0.76 and 0.75 respectively on the test dataset in classifying cardiovascular diseases. The methodology also outperformed three popular prior art approaches, when applied on the same dataset.

  9. AFSC/REFM: Seabird Necropsy dataset of North Pacific

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The seabird necropsy dataset contains information on seabird specimens that were collected under salvage and scientific collection permits primarily by...

  10. Dataset definition for CMS operations and physics analyses

    CERN Document Server

    AUTHOR|(CDS)2051291

    2016-01-01

    Data recorded at the CMS experiment are funnelled into streams, integrated in the HLT menu, and further organised in a hierarchical structure of primary datasets, secondary datasets, and dedicated skims. Datasets are defined according to the final-state particles reconstructed by the high level trigger, the data format and the use case (physics analysis, alignment and calibration, performance studies). During the first LHC run, new workflows have been added to this canonical scheme, to exploit at best the flexibility of the CMS trigger and data acquisition systems. The concept of data parking and data scouting have been introduced to extend the physics reach of CMS, offering the opportunity of defining physics triggers with extremely loose selections (e.g. dijet resonance trigger collecting data at a 1 kHz). In this presentation, we review the evolution of the dataset definition during the first run, and we discuss the plans for the second LHC run.

  11. USGS National Boundary Dataset (NBD) Downloadable Data Collection

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — The USGS Governmental Unit Boundaries dataset from The National Map (TNM) represents major civil areas for the Nation, including States or Territories, counties (or...

  12. Environmental Dataset Gateway (EDG) CS-W Interface

    Data.gov (United States)

    U.S. Environmental Protection Agency — Use the Environmental Dataset Gateway (EDG) to find and access EPA's environmental resources. Many options are available for easily reusing EDG content in other...

  13. Global Man-made Impervious Surface (GMIS) Dataset From Landsat

    Data.gov (United States)

    National Aeronautics and Space Administration — The Global Man-made Impervious Surface (GMIS) Dataset From Landsat consists of global estimates of fractional impervious cover derived from the Global Land Survey...

  14. A Comparative Analysis of Classification Algorithms on Diverse Datasets

    Directory of Open Access Journals (Sweden)

    M. Alghobiri

    2018-04-01

    Full Text Available Data mining involves the computational process to find patterns from large data sets. Classification, one of the main domains of data mining, involves known structure generalizing to apply to a new dataset and predict its class. There are various classification algorithms being used to classify various data sets. They are based on different methods such as probability, decision tree, neural network, nearest neighbor, boolean and fuzzy logic, kernel-based etc. In this paper, we apply three diverse classification algorithms on ten datasets. The datasets have been selected based on their size and/or number and nature of attributes. Results have been discussed using some performance evaluation measures like precision, accuracy, F-measure, Kappa statistics, mean absolute error, relative absolute error, ROC Area etc. Comparative analysis has been carried out using the performance evaluation measures of accuracy, precision, and F-measure. We specify features and limitations of the classification algorithms for the diverse nature datasets.

  15. Newton SSANTA Dr Water using POU filters dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — This dataset contains information about all the features extracted from the raw data files, the formulas that were assigned to some of these features, and the...

  16. Estimating parameters for probabilistic linkage of privacy-preserved datasets.

    Science.gov (United States)

    Brown, Adrian P; Randall, Sean M; Ferrante, Anna M; Semmens, James B; Boyd, James H

    2017-07-10

    Probabilistic record linkage is a process used to bring together person-based records from within the same dataset (de-duplication) or from disparate datasets using pairwise comparisons and matching probabilities. The linkage strategy and associated match probabilities are often estimated through investigations into data quality and manual inspection. However, as privacy-preserved datasets comprise encrypted data, such methods are not possible. In this paper, we present a method for estimating the probabilities and threshold values for probabilistic privacy-preserved record linkage using Bloom filters. Our method was tested through a simulation study using synthetic data, followed by an application using real-world administrative data. Synthetic datasets were generated with error rates from zero to 20% error. Our method was used to estimate parameters (probabilities and thresholds) for de-duplication linkages. Linkage quality was determined by F-measure. Each dataset was privacy-preserved using separate Bloom filters for each field. Match probabilities were estimated using the expectation-maximisation (EM) algorithm on the privacy-preserved data. Threshold cut-off values were determined by an extension to the EM algorithm allowing linkage quality to be estimated for each possible threshold. De-duplication linkages of each privacy-preserved dataset were performed using both estimated and calculated probabilities. Linkage quality using the F-measure at the estimated threshold values was also compared to the highest F-measure. Three large administrative datasets were used to demonstrate the applicability of the probability and threshold estimation technique on real-world data. Linkage of the synthetic datasets using the estimated probabilities produced an F-measure that was comparable to the F-measure using calculated probabilities, even with up to 20% error. Linkage of the administrative datasets using estimated probabilities produced an F-measure that was higher

  17. Toward computational cumulative biology by combining models of biological datasets.

    Science.gov (United States)

    Faisal, Ali; Peltonen, Jaakko; Georgii, Elisabeth; Rung, Johan; Kaski, Samuel

    2014-01-01

    A main challenge of data-driven sciences is how to make maximal use of the progressively expanding databases of experimental datasets in order to keep research cumulative. We introduce the idea of a modeling-based dataset retrieval engine designed for relating a researcher's experimental dataset to earlier work in the field. The search is (i) data-driven to enable new findings, going beyond the state of the art of keyword searches in annotations, (ii) modeling-driven, to include both biological knowledge and insights learned from data, and (iii) scalable, as it is accomplished without building one unified grand model of all data. Assuming each dataset has been modeled beforehand, by the researchers or automatically by database managers, we apply a rapidly computable and optimizable combination model to decompose a new dataset into contributions from earlier relevant models. By using the data-driven decomposition, we identify a network of interrelated datasets from a large annotated human gene expression atlas. While tissue type and disease were major driving forces for determining relevant datasets, the found relationships were richer, and the model-based search was more accurate than the keyword search; moreover, it recovered biologically meaningful relationships that are not straightforwardly visible from annotations-for instance, between cells in different developmental stages such as thymocytes and T-cells. Data-driven links and citations matched to a large extent; the data-driven links even uncovered corrections to the publication data, as two of the most linked datasets were not highly cited and turned out to have wrong publication entries in the database.

  18. Testing the Neutral Theory of Biodiversity with Human Microbiome Datasets

    OpenAIRE

    Li, Lianwei; Ma, Zhanshan (Sam)

    2016-01-01

    The human microbiome project (HMP) has made it possible to test important ecological theories for arguably the most important ecosystem to human health?the human microbiome. Existing limited number of studies have reported conflicting evidence in the case of the neutral theory; the present study aims to comprehensively test the neutral theory with extensive HMP datasets covering all five major body sites inhabited by the human microbiome. Utilizing 7437 datasets of bacterial community samples...

  19. General Purpose Multimedia Dataset - GarageBand 2008

    DEFF Research Database (Denmark)

    Meng, Anders

    This document describes a general purpose multimedia data-set to be used in cross-media machine learning problems. In more detail we describe the genre taxonomy applied at http://www.garageband.com, from where the data-set was collected, and how the taxonomy have been fused into a more human...... understandable taxonomy. Finally, a description of various features extracted from both the audio and text are presented....

  20. Artificial intelligence (AI) systems for interpreting complex medical datasets.

    Science.gov (United States)

    Altman, R B

    2017-05-01

    Advances in machine intelligence have created powerful capabilities in algorithms that find hidden patterns in data, classify objects based on their measured characteristics, and associate similar patients/diseases/drugs based on common features. However, artificial intelligence (AI) applications in medical data have several technical challenges: complex and heterogeneous datasets, noisy medical datasets, and explaining their output to users. There are also social challenges related to intellectual property, data provenance, regulatory issues, economics, and liability. © 2017 ASCPT.

  1. A Dataset for Three-Dimensional Distribution of 39 Elements Including Plant Nutrients and Other Metals and Metalloids in the Soils of a Forested Headwater Catchment.

    Science.gov (United States)

    Wu, B; Wiekenkamp, I; Sun, Y; Fisher, A S; Clough, R; Gottselig, N; Bogena, H; Pütz, T; Brüggemann, N; Vereecken, H; Bol, R

    2017-11-01

    Quantification and evaluation of elemental distribution in forested ecosystems are key requirements to understand element fluxes and their relationship with hydrological and biogeochemical processes in the system. However, datasets supporting such a study on the catchment scale are still limited. Here we provide a dataset comprising spatially highly resolved distributions of 39 elements in soil profiles of a small forested headwater catchment in western Germany () to gain a holistic picture of the state and fluxes of elements in the catchment. The elements include both plant nutrients and other metals and metalloids that were predominately derived from lithospheric or anthropogenic inputs, thereby allowing us to not only capture the nutrient status of the catchment but to also estimate the functional development of the ecosystem. Soil samples were collected at high lateral resolution (≤60 m), and element concentrations were determined vertically for four soil horizons (L/Of, Oh, A, B). From this, a three-dimensional view of the distribution of these elements could be established with high spatial resolution on the catchment scale in a temperate natural forested ecosystem. The dataset can be combined with other datasets and studies of the TERENO (Terrestrial Environmental Observatories) Data Discovery Portal () to reveal elemental fluxes, establish relations between elements and other soil properties, and/or as input for modeling elemental cycling in temperate forested ecosystems. Copyright © by the American Society of Agronomy, Crop Science Society of America, and Soil Science Society of America, Inc.

  2. Fast Flux Test Facility

    International Nuclear Information System (INIS)

    Munn, W.I.

    1981-01-01

    The Fast Flux Test Facility (FFTF), located on the Hanford site a few miles north of Richland, Washington, is a major link in the chain of development required to sustain and advance Liquid Metal Fast Breeder Reactor (LMFBR) technology in the United States. This 400 MWt sodium cooled reactor is a three loop design, is operated by Westinghouse Hanford Company for the US Department of Energy, and is the largest research reactor of its kind in the world. The purpose of the facility is three-fold: (1) to provide a test bed for components, materials, and breeder reactor fuels which can significantly extend resource reserves; (2) to produce a complete body of base data for the use of liquid sodium in heat transfer systens; and (3) to demonstrate inherent safety characteristics of LMFBR designs

  3. Heuristics for Relevancy Ranking of Earth Dataset Search Results

    Science.gov (United States)

    Lynnes, Christopher; Quinn, Patrick; Norton, James

    2016-01-01

    As the Variety of Earth science datasets increases, science researchers find it more challenging to discover and select the datasets that best fit their needs. The most common way of search providers to address this problem is to rank the datasets returned for a query by their likely relevance to the user. Large web page search engines typically use text matching supplemented with reverse link counts, semantic annotations and user intent modeling. However, this produces uneven results when applied to dataset metadata records simply externalized as a web page. Fortunately, data and search provides have decades of experience in serving data user communities, allowing them to form heuristics that leverage the structure in the metadata together with knowledge about the user community. Some of these heuristics include specific ways of matching the user input to the essential measurements in the dataset and determining overlaps of time range and spatial areas. Heuristics based on the novelty of the datasets can prioritize later, better versions of data over similar predecessors. And knowledge of how different user types and communities use data can be brought to bear in cases where characteristics of the user (discipline, expertise) or their intent (applications, research) can be divined. The Earth Observing System Data and Information System has begun implementing some of these heuristics in the relevancy algorithm of its Common Metadata Repository search engine.

  4. Derivation of ozone flux-yield relationships for lettuce: A key horticultural crop

    Energy Technology Data Exchange (ETDEWEB)

    Goumenaki, Eleni [Environmental and Molecular Plant Physiology, Institute for Research on the Environment and Sustainability, School of Biology and Psychology, Division of Biology, Devonshire Building, Newcastle University, Newcastle Upon Tyne NE1 7RU (United Kingdom); School of Agricultural Technology, Technological Education Institute of Crete, P.O. Box 1939, 71004 Heraklion (Greece); Fernandez, Ignacio Gonzalez [Environmental and Molecular Plant Physiology, Institute for Research on the Environment and Sustainability, School of Biology and Psychology, Division of Biology, Devonshire Building, Newcastle University, Newcastle Upon Tyne NE1 7RU (United Kingdom); CIEMAT, Ecotoxicology of Air Pollution, Avda. Complutense 22, 28040 Madrid (Spain); Papanikolaou, Antigoni [School of Agricultural Technology, Technological Education Institute of Crete, P.O. Box 1939, 71004 Heraklion (Greece); Papadopoulou, Despoina [School of Agricultural Technology, Technological Education Institute of Crete, P.O. Box 1939, 71004 Heraklion (Greece); Askianakis, Christos [School of Agricultural Technology, Technological Education Institute of Crete, P.O. Box 1939, 71004 Heraklion (Greece); Kouvarakis, George [Environmental and Chemical Processes Laboratory, Department of Chemistry, University of Crete, P.O. Box 1470, 71409 Heraklion (Greece); Barnes, Jeremy [Environmental and Molecular Plant Physiology, Institute for Research on the Environment and Sustainability, School of Biology and Psychology, Division of Biology, Devonshire Building, Newcastle University, Newcastle Upon Tyne NE1 7RU (United Kingdom)]. E-mail: j.d.barnes@ncl.ac.uk

    2007-04-15

    Ozone flux-response relationships were derived for lettuce, employing a multiplicative approach to model the manner in which stomatal conductance is influenced by key environmental variables, using a dataset collected during field experimentation in Crete and yield-response relationships derived from parallel open-top chamber experiments. Regional agronomic practices were adopted throughout. Computed versus measured data revealed that the derived model explained 51% (P < 0.001) of the observed variation in stomatal conductance. Concentration-based indices were compared with flux-based indices. Analyses revealed a significant relationship between accumulated stomatal ozone flux and yield employing flux threshold cut-offs up to 4 nmol m{sup -2} s{sup -1}. Regressions employing very low or zero flux thresholds resulted in the strongest yield-flux relationships (explaining {approx}80% (P < 0.05) of the variation in the dataset). - Establishment of ozone flux-yield relationships for a commercially-important horticultural crop grown widely in the Mediterranean.

  5. Flux compactifications and generalized geometries

    International Nuclear Information System (INIS)

    Grana, Mariana

    2006-01-01

    Following the lectures given at CERN Winter School 2006, we present a pedagogical overview of flux compactifications and generalized geometries, concentrating on closed string fluxes in type II theories. We start by reviewing the supersymmetric flux configurations with maximally symmetric four-dimensional spaces. We then discuss the no-go theorems (and their evasion) for compactifications with fluxes. We analyse the resulting four-dimensional effective theories for Calabi-Yau and Calabi-Yau orientifold compactifications, concentrating on the flux-induced superpotentials. We discuss the generic mechanism of moduli stabilization and illustrate with two examples: the conifold in IIB and a T 6 /(Z 3 x Z 3 ) torus in IIA. We finish by studying the effective action and flux vacua for generalized geometries in the context of generalized complex geometry

  6. Flux compactifications and generalized geometries

    Energy Technology Data Exchange (ETDEWEB)

    Grana, Mariana [Service de Physique Theorique, CEA/Saclay, 91191 Gif-sur-Yvette Cedex (France)

    2006-11-07

    Following the lectures given at CERN Winter School 2006, we present a pedagogical overview of flux compactifications and generalized geometries, concentrating on closed string fluxes in type II theories. We start by reviewing the supersymmetric flux configurations with maximally symmetric four-dimensional spaces. We then discuss the no-go theorems (and their evasion) for compactifications with fluxes. We analyse the resulting four-dimensional effective theories for Calabi-Yau and Calabi-Yau orientifold compactifications, concentrating on the flux-induced superpotentials. We discuss the generic mechanism of moduli stabilization and illustrate with two examples: the conifold in IIB and a T{sup 6} /(Z{sub 3} x Z{sub 3}) torus in IIA. We finish by studying the effective action and flux vacua for generalized geometries in the context of generalized complex geometry.

  7. 1km Global Terrestrial Carbon Flux: Estimations and Evaluations

    Science.gov (United States)

    Murakami, K.; Sasai, T.; Kato, S.; Saito, M.; Matsunaga, T.; Hiraki, K.; Maksyutov, S. S.

    2017-12-01

    Estimating global scale of the terrestrial carbon flux change with high accuracy and high resolution is important to understand global environmental changes. Furthermore the estimations of the global spatiotemporal distribution may contribute to the political and social activities such as REDD+. In order to reveal the current state of terrestrial carbon fluxes covering all over the world and a decadal scale. The satellite-based diagnostic biosphere model is suitable for achieving this purpose owing to observing on the present global land surface condition uniformly at some time interval. In this study, we estimated the global terrestrial carbon fluxes with 1km grids by using the terrestrial biosphere model (BEAMS). And we evaluated our new carbon flux estimations on various spatial scales and showed the transition of forest carbon stocks in some regions. Because BEAMS required high resolution meteorological data and satellite data as input data, we made 1km interpolated data using a kriging method. The data used in this study were JRA-55, GPCP, GOSAT L4B atmospheric CO2 data as meteorological data, and MODIS land product as land surface satellite data. Interpolating process was performed on the meteorological data because of insufficient resolution, but not on MODIS data. We evaluated our new carbon flux estimations using the flux tower measurement (FLUXNET2015 Datasets) in a point scale. We used 166 sites data for evaluating our model results. These flux sites are classified following vegetation type (DBF, EBF, ENF, mixed forests, grass lands, croplands, shrub lands, Savannas, wetlands). In global scale, the BEAMS estimations was underestimated compared to the flux measurements in the case of carbon uptake and release. The monthly variations of NEP showed relatively high correlations in DBF and mixed forests, but the correlation coefficients of EBF, ENF, and grass lands were less than 0.5. In the meteorological factors, air temperature and solar radiation showed

  8. SatelliteDL: a Toolkit for Analysis of Heterogeneous Satellite Datasets

    Science.gov (United States)

    Galloy, M. D.; Fillmore, D.

    2014-12-01

    SatelliteDL is an IDL toolkit for the analysis of satellite Earth observations from a diverse set of platforms and sensors. The core function of the toolkit is the spatial and temporal alignment of satellite swath and geostationary data. The design features an abstraction layer that allows for easy inclusion of new datasets in a modular way. Our overarching objective is to create utilities that automate the mundane aspects of satellite data analysis, are extensible and maintainable, and do not place limitations on the analysis itself. IDL has a powerful suite of statistical and visualization tools that can be used in conjunction with SatelliteDL. Toward this end we have constructed SatelliteDL to include (1) HTML and LaTeX API document generation,(2) a unit test framework,(3) automatic message and error logs,(4) HTML and LaTeX plot and table generation, and(5) several real world examples with bundled datasets available for download. For ease of use, datasets, variables and optional workflows may be specified in a flexible format configuration file. Configuration statements may specify, for example, a region and date range, and the creation of images, plots and statistical summary tables for a long list of variables. SatelliteDL enforces data provenance; all data should be traceable and reproducible. The output NetCDF file metadata holds a complete history of the original datasets and their transformations, and a method exists to reconstruct a configuration file from this information. Release 0.1.0 distributes with ingest methods for GOES, MODIS, VIIRS and CERES radiance data (L1) as well as select 2D atmosphere products (L2) such as aerosol and cloud (MODIS and VIIRS) and radiant flux (CERES). Future releases will provide ingest methods for ocean and land surface products, gridded and time averaged datasets (L3 Daily, Monthly and Yearly), and support for 3D products such as temperature and water vapor profiles. Emphasis will be on NPP Sensor, Environmental and

  9. Mining and Utilizing Dataset Relevancy from Oceanographic Dataset Metadata, Usage Metrics, and User Feedback to Improve Data Discovery and Access

    Data.gov (United States)

    National Aeronautics and Space Administration — We propose to mine and utilize the combination of Earth Science dataset, metadata with usage metrics and user feedback to objectively extract relevance for improved...

  10. Heat Flux Instrumentation Laboratory (HFIL)

    Data.gov (United States)

    Federal Laboratory Consortium — Description: The Heat Flux Instrumentation Laboratory is used to develop advanced, flexible, thin film gauge instrumentation for the Air Force Research Laboratory....

  11. Overview of NASA's Carbon Monitoring System Flux-Pilot Project

    Science.gov (United States)

    Pawson, Steven; Gunson, Michael R.; Jucks, Kenneth

    2011-01-01

    NASA's space-based observations of physical, chemical and biological parameters in the Earth System along with state-of-the-art modeling capabilities provide unique capabilities for analyses of the carbon cycle. The Carbon Monitoring System is developing an exploratory framework for detecting carbon in the environment and its changes, with a view towards contributing to national and international monitoring activities. The Flux-Pilot Project aims to provide a unified view of land-atmosphere and ocean-atmosphere carbon exchange, using observation-constrained models. Central to the project is the application of NASA's satellite observations (especially MODIS), the ACOS retrievals of the JAXA-GOSAT observations, and the "MERRA" meteorological reanalysis produced with GEOS-S. With a primary objective of estimating uncertainty in computed fluxes, two land- and two ocean-systems are run for 2009-2010 and compared with existing flux estimates. An transport model is used to evaluate simulated CO2 concentrations with in-situ and space-based observations, in order to assess the realism of the fluxes and how uncertainties in fluxes propagate into atmospheric concentrations that can be more readily evaluated. Finally, the atmospheric partial CO2 columns observed from space are inverted to give new estimates of surface fluxes, which are evaluated using the bottom-up estimates and independent datasets. The focus of this presentation will be on the science goals and current achievements of the pilot project, with emphasis on how policy-relevant questions help focus the scientific direction. Examples include the issue of what spatio-temporal resolution of fluxes can be detected from polar-orbiting satellites and whether it is possible to use space-based observations to separate contributions to atmospheric concentrations of (say) fossil-fuel and biological activity

  12. KoFlux: Korean Regional Flux Network in AsiaFlux

    Science.gov (United States)

    Kim, J.

    2002-12-01

    AsiaFlux, the Asian arm of FLUXNET, held the Second International Workshop on Advanced Flux Network and Flux Evaluation in Jeju Island, Korea on 9-11 January 2002. In order to facilitate comprehensive Asia-wide studies of ecosystem fluxes, the meeting launched KoFlux, a new Korean regional network of long-term micrometeorological flux sites. For a successful assessment of carbon exchange between terrestrial ecosystems and the atmosphere, an accurate measurement of surface fluxes of energy and water is one of the prerequisites. During the 7th Global Energy and Water Cycle Experiment (GEWEX) Asian Monsoon Experiment (GAME) held in Nagoya, Japan on 1-2 October 2001, the Implementation Committee of the Coordinated Enhanced Observing Period (CEOP) was established. One of the immediate tasks of CEOP was and is to identify the reference sites to monitor energy and water fluxes over the Asian continent. Subsequently, to advance the regional and global network of these reference sites in the context of both FLUXNET and CEOP, the Korean flux community has re-organized the available resources to establish a new regional network, KoFlux. We have built up domestic network sites (equipped with wind profiler and radiosonde measurements) over deciduous and coniferous forests, urban and rural rice paddies and coastal farmland. As an outreach through collaborations with research groups in Japan, China and Thailand, we also proposed international flux sites at ecologically and climatologically important locations such as a prairie on the Tibetan plateau, tropical forest with mixed and rapid land use change in northern Thailand. Several sites in KoFlux already begun to accumulate interesting data and some highlights are presented at the meeting. The sciences generated by flux networks in other continents have proven the worthiness of a global array of micrometeorological flux towers. It is our intent that the launch of KoFlux would encourage other scientists to initiate and

  13. EEG datasets for motor imagery brain-computer interface.

    Science.gov (United States)

    Cho, Hohyun; Ahn, Minkyu; Ahn, Sangtae; Kwon, Moonyoung; Jun, Sung Chan

    2017-07-01

    Most investigators of brain-computer interface (BCI) research believe that BCI can be achieved through induced neuronal activity from the cortex, but not by evoked neuronal activity. Motor imagery (MI)-based BCI is one of the standard concepts of BCI, in that the user can generate induced activity by imagining motor movements. However, variations in performance over sessions and subjects are too severe to overcome easily; therefore, a basic understanding and investigation of BCI performance variation is necessary to find critical evidence of performance variation. Here we present not only EEG datasets for MI BCI from 52 subjects, but also the results of a psychological and physiological questionnaire, EMG datasets, the locations of 3D EEG electrodes, and EEGs for non-task-related states. We validated our EEG datasets by using the percentage of bad trials, event-related desynchronization/synchronization (ERD/ERS) analysis, and classification analysis. After conventional rejection of bad trials, we showed contralateral ERD and ipsilateral ERS in the somatosensory area, which are well-known patterns of MI. Finally, we showed that 73.08% of datasets (38 subjects) included reasonably discriminative information. Our EEG datasets included the information necessary to determine statistical significance; they consisted of well-discriminated datasets (38 subjects) and less-discriminative datasets. These may provide researchers with opportunities to investigate human factors related to MI BCI performance variation, and may also achieve subject-to-subject transfer by using metadata, including a questionnaire, EEG coordinates, and EEGs for non-task-related states. © The Authors 2017. Published by Oxford University Press.

  14. From elementary flux modes to elementary flux vectors: Metabolic pathway analysis with arbitrary linear flux constraints

    Science.gov (United States)

    Klamt, Steffen; Gerstl, Matthias P.; Jungreuthmayer, Christian; Mahadevan, Radhakrishnan; Müller, Stefan

    2017-01-01

    Elementary flux modes (EFMs) emerged as a formal concept to describe metabolic pathways and have become an established tool for constraint-based modeling and metabolic network analysis. EFMs are characteristic (support-minimal) vectors of the flux cone that contains all feasible steady-state flux vectors of a given metabolic network. EFMs account for (homogeneous) linear constraints arising from reaction irreversibilities and the assumption of steady state; however, other (inhomogeneous) linear constraints, such as minimal and maximal reaction rates frequently used by other constraint-based techniques (such as flux balance analysis [FBA]), cannot be directly integrated. These additional constraints further restrict the space of feasible flux vectors and turn the flux cone into a general flux polyhedron in which the concept of EFMs is not directly applicable anymore. For this reason, there has been a conceptual gap between EFM-based (pathway) analysis methods and linear optimization (FBA) techniques, as they operate on different geometric objects. One approach to overcome these limitations was proposed ten years ago and is based on the concept of elementary flux vectors (EFVs). Only recently has the community started to recognize the potential of EFVs for metabolic network analysis. In fact, EFVs exactly represent the conceptual development required to generalize the idea of EFMs from flux cones to flux polyhedra. This work aims to present a concise theoretical and practical introduction to EFVs that is accessible to a broad audience. We highlight the close relationship between EFMs and EFVs and demonstrate that almost all applications of EFMs (in flux cones) are possible for EFVs (in flux polyhedra) as well. In fact, certain properties can only be studied with EFVs. Thus, we conclude that EFVs provide a powerful and unifying framework for constraint-based modeling of metabolic networks. PMID:28406903

  15. Comparison of CORA and EN4 in-situ datasets validation methods, toward a better quality merged dataset.

    Science.gov (United States)

    Szekely, Tanguy; Killick, Rachel; Gourrion, Jerome; Reverdin, Gilles

    2017-04-01

    CORA and EN4 are both global delayed time mode validated in-situ ocean temperature and salinity datasets distributed by the Met Office (http://www.metoffice.gov.uk/) and Copernicus (www.marine.copernicus.eu). A large part of the profiles distributed by CORA and EN4 in recent years are Argo profiles from the ARGO DAC, but profiles are also extracted from the World Ocean Database and TESAC profiles from GTSPP. In the case of CORA, data coming from the EUROGOOS Regional operationnal oserving system( ROOS) operated by European institutes no managed by National Data Centres and other datasets of profiles povided by scientific sources can also be found (Sea mammals profiles from MEOP, XBT datasets from cruises ...). (EN4 also takes data from the ASBO dataset to supplement observations in the Arctic). First advantage of this new merge product is to enhance the space and time coverage at global and european scales for the period covering 1950 till a year before the current year. This product is updated once a year and T&S gridded fields are alos generated for the period 1990-year n-1. The enhancement compared to the revious CORA product will be presented Despite the fact that the profiles distributed by both datasets are mostly the same, the quality control procedures developed by the Met Office and Copernicus teams differ, sometimes leading to different quality control flags for the same profile. Started in 2016 a new study started that aims to compare both validation procedures to move towards a Copernicus Marine Service dataset with the best features of CORA and EN4 validation.A reference data set composed of the full set of in-situ temperature and salinity measurements collected by Coriolis during 2015 is used. These measurements have been made thanks to wide range of instruments (XBTs, CTDs, Argo floats, Instrumented sea mammals,...), covering the global ocean. The reference dataset has been validated simultaneously by both teams.An exhaustive comparison of the

  16. Utilizing LiDAR Datasets From Experimental Watersheds to Advance Ecohydrological Understanding in Seasonally Snow-Covered Forests

    Science.gov (United States)

    Harpold, A. A.; Broxton, P. D.; Guo, Q.; Barlage, M. J.; Gochis, D. J.

    2014-12-01

    The Western U.S. is strongly reliant on snowmelt from forested areas for ecosystem services and downstream populations. The ability to manage water resources from snow-covered forests faces major challenges from drought, disturbance, and regional changes in climate. An exciting avenue for improving ecohydrological process understanding is Light Detection and Ranging (LiDAR) because the technology simultaneously observes topography, forest properties, and snow/ice at high-resolution (100 km2). The availability and quality of LiDAR datasets is increasing rapidly, however they remain under-utilized for process-based ecohydrology investigations. This presentation will illustrate how LiDAR datasets from the Critical Zone Observatory (CZO) network have been applied to advance ecohydrological understanding through direct empirical analysis, as well as model parameterization and verification. Direct analysis of the datasets has proved fruitful for pre- and post-disturbance snow distribution estimates and interpreting in-situ snow depth measurements across sites. In addition, we illustrate the potential value of LiDAR to parameterize and verify of physical models with two examples. First, we use LiDAR to parameterize a land surface model, Noah multi-parameterization (Noah-MP), to investigate the sensitivity of modeled water and energy fluxes to high-resolution forest information. Second, we present a Snow Physics and Laser Mapping (SnowPALM) model that is parameterized with LiDAR information at its native 1-m scale. Both modeling studies demonstrate the value of LiDAR for representing processes with greater fidelity. More importantly, the increased model fidelity led to different estimates of water and energy fluxes at larger, watershed scales. Creating a network of experimental watersheds with LiDAR datasets offers the potential to test theories and models in previously unexplored ways.

  17. Wind and wave dataset for Matara, Sri Lanka

    Science.gov (United States)

    Luo, Yao; Wang, Dongxiao; Priyadarshana Gamage, Tilak; Zhou, Fenghua; Madusanka Widanage, Charith; Liu, Taiwei

    2018-01-01

    We present a continuous in situ hydro-meteorology observational dataset from a set of instruments first deployed in December 2012 in the south of Sri Lanka, facing toward the north Indian Ocean. In these waters, simultaneous records of wind and wave data are sparse due to difficulties in deploying measurement instruments, although the area hosts one of the busiest shipping lanes in the world. This study describes the survey, deployment, and measurements of wind and waves, with the aim of offering future users of the dataset the most comprehensive and as much information as possible. This dataset advances our understanding of the nearshore hydrodynamic processes and wave climate, including sea waves and swells, in the north Indian Ocean. Moreover, it is a valuable resource for ocean model parameterization and validation. The archived dataset (Table 1) is examined in detail, including wave data at two locations with water depths of 20 and 10 m comprising synchronous time series of wind, ocean astronomical tide, air pressure, etc. In addition, we use these wave observations to evaluate the ERA-Interim reanalysis product. Based on Buoy 2 data, the swells are the main component of waves year-round, although monsoons can markedly alter the proportion between swell and wind sea. The dataset (Luo et al., 2017) is publicly available from Science Data Bank (https://doi.org/10.11922/sciencedb.447).

  18. The LANDFIRE Refresh strategy: updating the national dataset

    Science.gov (United States)

    Nelson, Kurtis J.; Connot, Joel A.; Peterson, Birgit E.; Martin, Charley

    2013-01-01

    The LANDFIRE Program provides comprehensive vegetation and fuel datasets for the entire United States. As with many large-scale ecological datasets, vegetation and landscape conditions must be updated periodically to account for disturbances, growth, and natural succession. The LANDFIRE Refresh effort was the first attempt to consistently update these products nationwide. It incorporated a combination of specific systematic improvements to the original LANDFIRE National data, remote sensing based disturbance detection methods, field collected disturbance information, vegetation growth and succession modeling, and vegetation transition processes. This resulted in the creation of two complete datasets for all 50 states: LANDFIRE Refresh 2001, which includes the systematic improvements, and LANDFIRE Refresh 2008, which includes the disturbance and succession updates to the vegetation and fuel data. The new datasets are comparable for studying landscape changes in vegetation type and structure over a decadal period, and provide the most recent characterization of fuel conditions across the country. The applicability of the new layers is discussed and the effects of using the new fuel datasets are demonstrated through a fire behavior modeling exercise using the 2011 Wallow Fire in eastern Arizona as an example.

  19. Interactive visualization and analysis of multimodal datasets for surgical applications.

    Science.gov (United States)

    Kirmizibayrak, Can; Yim, Yeny; Wakid, Mike; Hahn, James

    2012-12-01

    Surgeons use information from multiple sources when making surgical decisions. These include volumetric datasets (such as CT, PET, MRI, and their variants), 2D datasets (such as endoscopic videos), and vector-valued datasets (such as computer simulations). Presenting all the information to the user in an effective manner is a challenging problem. In this paper, we present a visualization approach that displays the information from various sources in a single coherent view. The system allows the user to explore and manipulate volumetric datasets, display analysis of dataset values in local regions, combine 2D and 3D imaging modalities and display results of vector-based computer simulations. Several interaction methods are discussed: in addition to traditional interfaces including mouse and trackers, gesture-based natural interaction methods are shown to control these visualizations with real-time performance. An example of a medical application (medialization laryngoplasty) is presented to demonstrate how the combination of different modalities can be used in a surgical setting with our approach.

  20. Wind and wave dataset for Matara, Sri Lanka

    Directory of Open Access Journals (Sweden)

    Y. Luo

    2018-01-01

    Full Text Available We present a continuous in situ hydro-meteorology observational dataset from a set of instruments first deployed in December 2012 in the south of Sri Lanka, facing toward the north Indian Ocean. In these waters, simultaneous records of wind and wave data are sparse due to difficulties in deploying measurement instruments, although the area hosts one of the busiest shipping lanes in the world. This study describes the survey, deployment, and measurements of wind and waves, with the aim of offering future users of the dataset the most comprehensive and as much information as possible. This dataset advances our understanding of the nearshore hydrodynamic processes and wave climate, including sea waves and swells, in the north Indian Ocean. Moreover, it is a valuable resource for ocean model parameterization and validation. The archived dataset (Table 1 is examined in detail, including wave data at two locations with water depths of 20 and 10 m comprising synchronous time series of wind, ocean astronomical tide, air pressure, etc. In addition, we use these wave observations to evaluate the ERA-Interim reanalysis product. Based on Buoy 2 data, the swells are the main component of waves year-round, although monsoons can markedly alter the proportion between swell and wind sea. The dataset (Luo et al., 2017 is publicly available from Science Data Bank (https://doi.org/10.11922/sciencedb.447.

  1. Process mining in oncology using the MIMIC-III dataset

    Science.gov (United States)

    Prima Kurniati, Angelina; Hall, Geoff; Hogg, David; Johnson, Owen

    2018-03-01

    Process mining is a data analytics approach to discover and analyse process models based on the real activities captured in information systems. There is a growing body of literature on process mining in healthcare, including oncology, the study of cancer. In earlier work we found 37 peer-reviewed papers describing process mining research in oncology with a regular complaint being the limited availability and accessibility of datasets with suitable information for process mining. Publicly available datasets are one option and this paper describes the potential to use MIMIC-III, for process mining in oncology. MIMIC-III is a large open access dataset of de-identified patient records. There are 134 publications listed as using the MIMIC dataset, but none of them have used process mining. The MIMIC-III dataset has 16 event tables which are potentially useful for process mining and this paper demonstrates the opportunities to use MIMIC-III for process mining in oncology. Our research applied the L* lifecycle method to provide a worked example showing how process mining can be used to analyse cancer pathways. The results and data quality limitations are discussed along with opportunities for further work and reflection on the value of MIMIC-III for reproducible process mining research.

  2. Flux trapping in superconducting cavities

    International Nuclear Information System (INIS)

    Vallet, C.; Bolore, M.; Bonin, B.; Charrier, J.P.; Daillant, B.; Gratadour, J.; Koechlin, F.; Safa, H.

    1992-01-01

    The flux trapped in various field cooled Nb and Pb samples has been measured. For ambient fields smaller than 3 Gauss, 100% of the flux is trapped. The consequences of this result on the behavior of superconducting RF cavities are discussed. (author) 12 refs.; 2 figs

  3. Squeezing Flux Out of Fat

    DEFF Research Database (Denmark)

    Gonzalez-Franquesa, Alba; Patti, Mary-Elizabeth

    2018-01-01

    Merging transcriptomics or metabolomics data remains insufficient for metabolic flux estimation. Ramirez et al. integrate a genome-scale metabolic model with extracellular flux data to predict and validate metabolic differences between white and brown adipose tissue. This method allows both metab...

  4. Data Acquisition and Flux Calculations

    DEFF Research Database (Denmark)

    Rebmann, C.; Kolle, O; Heinesch, B

    2012-01-01

    In this chapter, the basic theory and the procedures used to obtain turbulent fluxes of energy, mass, and momentum with the eddy covariance technique will be detailed. This includes a description of data acquisition, pretreatment of high-frequency data and flux calculation....

  5. A Spatial-Temporal Comparison of Lake Mendota CO2 Fluxes and Collection Methods

    Science.gov (United States)

    Baldocchi, A. K.; Reed, D. E.; Desai, A. R.; Loken, L. C.; Schramm, P.; Stanley, E. H.

    2017-12-01

    Monitoring of carbon fluxes at the lake/atmosphere interface can help us determine baselines from which to understand responses in both space and time that may result from our warming climate or increasing nutrient inputs. Since recent research has shown lakes to be hotspots of global carbon cycling, it is important to quantify carbon sink and source dynamics as well as to verify observations between multiple methods in the context of long-term data collection efforts. Here we evaluate a new method for measuring space and time variation in CO2 fluxes based on novel speedboat-based collection method of aquatic greenhouse gas concentrations and a flux computation and interpolation algorithm. Two-hundred and forty-nine consecutive days of spatial flux maps over the 2016 open ice period were compared to ongoing eddy covariance tower flux measurements on the shore of Lake Mendota, Wisconsin US using a flux footprint analysis. Spatial and temporal alignments of the fluxes from these two observational datasets revealed both similar trends from daily to seasonal timescales as well as biases between methods. For example, throughout the Spring carbon fluxes showed strong correlation although off by an order of magnitude. Isolating physical patterns of agreement between the two methods of the lake/atmosphere CO2 fluxes allows us to pinpoint where biology and physical drivers contribute to the global carbon cycle and help improve modelling of lakes and utilize lakes as leading indicators of climate change.

  6. How to choose methods for lake greenhouse gas flux measurements?

    Science.gov (United States)

    Bastviken, David

    2017-04-01

    Lake greenhouse gas (GHG) fluxes are increasingly recognized as important for lake ecosystems as well as for large scale carbon and GHG budgets. However, many of our flux estimates are uncertain and it can be discussed if the presently available data is representative for the systems studied or not. Data are also very limited for some important flux pathways. Hence, many ongoing efforts try to better constrain fluxes and understand flux regulation. A fundamental challenge towards improved knowledge and when starting new studies is what methods to choose. A variety of approaches to measure aquatic GHG exchange is used and data from different methods and methodological approaches have often been treated as equally valid to create large datasets for extrapolations and syntheses. However, data from different approaches may cover different flux pathways or spatio-temporal domains and are thus not always comparable. Method inter-comparisons and critical method evaluations addressing these issues are rare. Emerging efforts to organize systematic multi-lake monitoring networks for GHG fluxes leads to method choices that may set the foundation for decades of data generation and therefore require fundamental evaluation of different approaches. The method choices do not only regard the equipment but also for example consideration of overall measurement design and field approaches, relevant spatial and temporal resolution for different flux components, and accessory variables to measure. In addition, consideration of how to design monitoring approaches being affordable, suitable for widespread (global) use, and comparable across regions is needed. Inspired by discussions with Prof. Dr. Cristian Blodau during the EGU General Assembly 2016, this presentation aims to (1) illustrate fundamental pros and cons for a number of common methods, (2) show how common methodological approaches originally adapted for other environments can be improved for lake flux measurements, (3) suggest

  7. Solar proton fluxes since 1956

    International Nuclear Information System (INIS)

    Reedy, R.C.

    1977-01-01

    The fluxes of protons emitted during solar flares since 1956 were evaluated. The depth-versus-activity profiles of 56 Co in several lunar rocks are consistent with the solar-proton fluxes detected by experiments on several satellites. Only about 20% of the solar-proton-induced activities of 22 Na and 55 Fe in lunar rocks from early Apollo missions were produced by protons emitted from the sun during solar cycle 20 (1965--1975). The depth-versus-activity data for these radionuclides in several lunar rocks were used to determine the fluxes of protons during solar cycle 19 (1954--1964). The average proton fluxes for cycle 19 are about five times those for both the last million years and for cycle 20. These solar-proton flux variations correlate with changes in sunspot activity

  8. Continuous SO2 flux measurements for Vulcano Island, Italy

    Directory of Open Access Journals (Sweden)

    Fabio Vita

    2012-06-01

    Full Text Available The La Fossa cone of Vulcano Island (Aeolian Archipelago, Italy is a closed conduit volcano. Today, Vulcano Island is characterized by sulfataric activity, with a large fumarolic field that is mainly located in the summit area. A scanning differential optical absorption spectroscopy instrument designed by the Optical Sensing Group of Chalmers University of Technology in Göteborg, Sweden, was installed in the framework of the European project "Network for Observation of Volcanic and Atmospheric Change", in March 2008. This study presents the first dataset of SO2 plume fluxes recorded for a closed volcanic system. Between 2008 and 2010, the SO2 fluxes recorded showed average values of 12 t.d–1 during the normal sulfataric activity of Vulcano Island, with one exceptional event of strong degassing that occurred between September and December, 2009, when the SO2 emissions reached up to 100 t.d–1.

  9. Recent Development on the NOAA's Global Surface Temperature Dataset

    Science.gov (United States)

    Zhang, H. M.; Huang, B.; Boyer, T.; Lawrimore, J. H.; Menne, M. J.; Rennie, J.

    2016-12-01

    Global Surface Temperature (GST) is one of the most widely used indicators for climate trend and extreme analyses. A widely used GST dataset is the NOAA merged land-ocean surface temperature dataset known as NOAAGlobalTemp (formerly MLOST). The NOAAGlobalTemp had recently been updated from version 3.5.4 to version 4. The update includes a significant improvement in the ocean surface component (Extended Reconstructed Sea Surface Temperature or ERSST, from version 3b to version 4) which resulted in an increased temperature trends in recent decades. Since then, advancements in both the ocean component (ERSST) and land component (GHCN-Monthly) have been made, including the inclusion of Argo float SSTs and expanded EOT modes in ERSST, and the use of ISTI databank in GHCN-Monthly. In this presentation, we describe the impact of those improvements on the merged global temperature dataset, in terms of global trends and other aspects.

  10. Synthetic ALSPAC longitudinal datasets for the Big Data VR project.

    Science.gov (United States)

    Avraam, Demetris; Wilson, Rebecca C; Burton, Paul

    2017-01-01

    Three synthetic datasets - of observation size 15,000, 155,000 and 1,555,000 participants, respectively - were created by simulating eleven cardiac and anthropometric variables from nine collection ages of the ALSAPC birth cohort study. The synthetic datasets retain similar data properties to the ALSPAC study data they are simulated from (co-variance matrices, as well as the mean and variance values of the variables) without including the original data itself or disclosing participant information.  In this instance, the three synthetic datasets have been utilised in an academia-industry collaboration to build a prototype virtual reality data analysis software, but they could have a broader use in method and software development projects where sensitive data cannot be freely shared.

  11. The OXL format for the exchange of integrated datasets

    Directory of Open Access Journals (Sweden)

    Taubert Jan

    2007-12-01

    Full Text Available A prerequisite for systems biology is the integration and analysis of heterogeneous experimental data stored in hundreds of life-science databases and millions of scientific publications. Several standardised formats for the exchange of specific kinds of biological information exist. Such exchange languages facilitate the integration process; however they are not designed to transport integrated datasets. A format for exchanging integrated datasets needs to i cover data from a broad range of application domains, ii be flexible and extensible to combine many different complex data structures, iii include metadata and semantic definitions, iv include inferred information, v identify the original data source for integrated entities and vi transport large integrated datasets. Unfortunately, none of the exchange formats from the biological domain (e.g. BioPAX, MAGE-ML, PSI-MI, SBML or the generic approaches (RDF, OWL fulfil these requirements in a systematic way.

  12. Dataset of transcriptional landscape of B cell early activation

    Directory of Open Access Journals (Sweden)

    Alexander S. Garruss

    2015-09-01

    Full Text Available Signaling via B cell receptors (BCR and Toll-like receptors (TLRs result in activation of B cells with distinct physiological outcomes, but transcriptional regulatory mechanisms that drive activation and distinguish these pathways remain unknown. At early time points after BCR and TLR ligand exposure, 0.5 and 2 h, RNA-seq was performed allowing observations on rapid transcriptional changes. At 2 h, ChIP-seq was performed to allow observations on important regulatory mechanisms potentially driving transcriptional change. The dataset includes RNA-seq, ChIP-seq of control (Input, RNA Pol II, H3K4me3, H3K27me3, and a separate RNA-seq for miRNA expression, which can be found at Gene Expression Omnibus Dataset GSE61608. Here, we provide details on the experimental and analysis methods used to obtain and analyze this dataset and to examine the transcriptional landscape of B cell early activation.

  13. The Global Precipitation Climatology Project (GPCP) Combined Precipitation Dataset

    Science.gov (United States)

    Huffman, George J.; Adler, Robert F.; Arkin, Philip; Chang, Alfred; Ferraro, Ralph; Gruber, Arnold; Janowiak, John; McNab, Alan; Rudolf, Bruno; Schneider, Udo

    1997-01-01

    The Global Precipitation Climatology Project (GPCP) has released the GPCP Version 1 Combined Precipitation Data Set, a global, monthly precipitation dataset covering the period July 1987 through December 1995. The primary product in the dataset is a merged analysis incorporating precipitation estimates from low-orbit-satellite microwave data, geosynchronous-orbit -satellite infrared data, and rain gauge observations. The dataset also contains the individual input fields, a combination of the microwave and infrared satellite estimates, and error estimates for each field. The data are provided on 2.5 deg x 2.5 deg latitude-longitude global grids. Preliminary analyses show general agreement with prior studies of global precipitation and extends prior studies of El Nino-Southern Oscillation precipitation patterns. At the regional scale there are systematic differences with standard climatologies.

  14. A high-resolution European dataset for hydrologic modeling

    Science.gov (United States)

    Ntegeka, Victor; Salamon, Peter; Gomes, Goncalo; Sint, Hadewij; Lorini, Valerio; Thielen, Jutta

    2013-04-01

    There is an increasing demand for large scale hydrological models not only in the field of modeling the impact of climate change on water resources but also for disaster risk assessments and flood or drought early warning systems. These large scale models need to be calibrated and verified against large amounts of observations in order to judge their capabilities to predict the future. However, the creation of large scale datasets is challenging for it requires collection, harmonization, and quality checking of large amounts of observations. For this reason, only a limited number of such datasets exist. In this work, we present a pan European, high-resolution gridded dataset of meteorological observations (EFAS-Meteo) which was designed with the aim to drive a large scale hydrological model. Similar European and global gridded datasets already exist, such as the HadGHCND (Caesar et al., 2006), the JRC MARS-STAT database (van der Goot and Orlandi, 2003) and the E-OBS gridded dataset (Haylock et al., 2008). However, none of those provide similarly high spatial resolution and/or a complete set of variables to force a hydrologic model. EFAS-Meteo contains daily maps of precipitation, surface temperature (mean, minimum and maximum), wind speed and vapour pressure at a spatial grid resolution of 5 x 5 km for the time period 1 January 1990 - 31 December 2011. It furthermore contains calculated radiation, which is calculated by using a staggered approach depending on the availability of sunshine duration, cloud cover and minimum and maximum temperature, and evapotranspiration (potential evapotranspiration, bare soil and open water evapotranspiration). The potential evapotranspiration was calculated using the Penman-Monteith equation with the above-mentioned meteorological variables. The dataset was created as part of the development of the European Flood Awareness System (EFAS) and has been continuously updated throughout the last years. The dataset variables are used as

  15. Visualization of conserved structures by fusing highly variable datasets.

    Science.gov (United States)

    Silverstein, Jonathan C; Chhadia, Ankur; Dech, Fred

    2002-01-01

    Skill, effort, and time are required to identify and visualize anatomic structures in three-dimensions from radiological data. Fundamentally, automating these processes requires a technique that uses symbolic information not in the dynamic range of the voxel data. We were developing such a technique based on mutual information for automatic multi-modality image fusion (MIAMI Fuse, University of Michigan). This system previously demonstrated facility at fusing one voxel dataset with integrated symbolic structure information to a CT dataset (different scale and resolution) from the same person. The next step of development of our technique was aimed at accommodating the variability of anatomy from patient to patient by using warping to fuse our standard dataset to arbitrary patient CT datasets. A standard symbolic information dataset was created from the full color Visible Human Female by segmenting the liver parenchyma, portal veins, and hepatic veins and overwriting each set of voxels with a fixed color. Two arbitrarily selected patient CT scans of the abdomen were used for reference datasets. We used the warping functions in MIAMI Fuse to align the standard structure data to each patient scan. The key to successful fusion was the focused use of multiple warping control points that place themselves around the structure of interest automatically. The user assigns only a few initial control points to align the scans. Fusion 1 and 2 transformed the atlas with 27 points around the liver to CT1 and CT2 respectively. Fusion 3 transformed the atlas with 45 control points around the liver to CT1 and Fusion 4 transformed the atlas with 5 control points around the portal vein. The CT dataset is augmented with the transformed standard structure dataset, such that the warped structure masks are visualized in combination with the original patient dataset. This combined volume visualization is then rendered interactively in stereo on the ImmersaDesk in an immersive Virtual

  16. A cross-country Exchange Market Pressure (EMP) dataset.

    Science.gov (United States)

    Desai, Mohit; Patnaik, Ila; Felman, Joshua; Shah, Ajay

    2017-06-01

    The data presented in this article are related to the research article titled - "An exchange market pressure measure for cross country analysis" (Patnaik et al. [1]). In this article, we present the dataset for Exchange Market Pressure values (EMP) for 139 countries along with their conversion factors, ρ (rho). Exchange Market Pressure, expressed in percentage change in exchange rate, measures the change in exchange rate that would have taken place had the central bank not intervened. The conversion factor ρ can interpreted as the change in exchange rate associated with $1 billion of intervention. Estimates of conversion factor ρ allow us to calculate a monthly time series of EMP for 139 countries. Additionally, the dataset contains the 68% confidence interval (high and low values) for the point estimates of ρ 's. Using the standard errors of estimates of ρ 's, we obtain one sigma intervals around mean estimates of EMP values. These values are also reported in the dataset.

  17. Fractional flux excitations and flux creep in a superconducting film

    International Nuclear Information System (INIS)

    Lyuksyutov, I.F.

    1995-01-01

    We consider the transport properties of a modulated superconducting film in a magnetic field parallel to the film. Modulation can be either intrinsic, due to the layered structure of the high-T c superconductors, or artificial, e.g. due to thickness modulation. This system has an infinite set ( >) of pinned phases. In the pinned phase the excitation of flux loops with a fractional number of flux quanta by the applied current j results in flux creep with a generated voltage V ∝ exp[-jo/j[. (orig.)

  18. Dataset of herbarium specimens of threatened vascular plants in Catalonia.

    Science.gov (United States)

    Nualart, Neus; Ibáñez, Neus; Luque, Pere; Pedrol, Joan; Vilar, Lluís; Guàrdia, Roser

    2017-01-01

    This data paper describes a specimens' dataset of the Catalonian threatened vascular plants conserved in five public Catalonian herbaria (BC, BCN, HGI, HBIL and MTTE). Catalonia is an administrative region of Spain that includes large autochthon plants diversity and 199 taxa with IUCN threatened categories (EX, EW, RE, CR, EN and VU). This dataset includes 1,618 records collected from 17 th century to nowadays. For each specimen, the species name, locality indication, collection date, collector, ecology and revision label are recorded. More than 94% of the taxa are represented in the herbaria, which evidence the paper of the botanical collections as an essential source of occurrence data.

  19. A Large-Scale 3D Object Recognition dataset

    DEFF Research Database (Denmark)

    Sølund, Thomas; Glent Buch, Anders; Krüger, Norbert

    2016-01-01

    geometric groups; concave, convex, cylindrical and flat 3D object models. The object models have varying amount of local geometric features to challenge existing local shape feature descriptors in terms of descriptiveness and robustness. The dataset is validated in a benchmark which evaluates the matching...... performance of 7 different state-of-the-art local shape descriptors. Further, we validate the dataset in a 3D object recognition pipeline. Our benchmark shows as expected that local shape feature descriptors without any global point relation across the surface have a poor matching performance with flat...

  20. Traffic sign classification with dataset augmentation and convolutional neural network

    Science.gov (United States)

    Tang, Qing; Kurnianggoro, Laksono; Jo, Kang-Hyun

    2018-04-01

    This paper presents a method for traffic sign classification using a convolutional neural network (CNN). In this method, firstly we transfer a color image into grayscale, and then normalize it in the range (-1,1) as the preprocessing step. To increase robustness of classification model, we apply a dataset augmentation algorithm and create new images to train the model. To avoid overfitting, we utilize a dropout module before the last fully connection layer. To assess the performance of the proposed method, the German traffic sign recognition benchmark (GTSRB) dataset is utilized. Experimental results show that the method is effective in classifying traffic signs.

  1. Towards interoperable and reproducible QSAR analyses: Exchange of datasets.

    Science.gov (United States)

    Spjuth, Ola; Willighagen, Egon L; Guha, Rajarshi; Eklund, Martin; Wikberg, Jarl Es

    2010-06-30

    QSAR is a widely used method to relate chemical structures to responses or properties based on experimental observations. Much effort has been made to evaluate and validate the statistical modeling in QSAR, but these analyses treat the dataset as fixed. An overlooked but highly important issue is the validation of the setup of the dataset, which comprises addition of chemical structures as well as selection of descriptors and software implementations prior to calculations. This process is hampered by the lack of standards and exchange formats in the field, making it virtually impossible to reproduce and validate analyses and drastically constrain collaborations and re-use of data. We present a step towards standardizing QSAR analyses by defining interoperable and reproducible QSAR datasets, consisting of an open XML format (QSAR-ML) which builds on an open and extensible descriptor ontology. The ontology provides an extensible way of uniquely defining descriptors for use in QSAR experiments, and the exchange format supports multiple versioned implementations of these descriptors. Hence, a dataset described by QSAR-ML makes its setup completely reproducible. We also provide a reference implementation as a set of plugins for Bioclipse which simplifies setup of QSAR datasets, and allows for exporting in QSAR-ML as well as old-fashioned CSV formats. The implementation facilitates addition of new descriptor implementations from locally installed software and remote Web services; the latter is demonstrated with REST and XMPP Web services. Standardized QSAR datasets open up new ways to store, query, and exchange data for subsequent analyses. QSAR-ML supports completely reproducible creation of datasets, solving the problems of defining which software components were used and their versions, and the descriptor ontology eliminates confusions regarding descriptors by defining them crisply. This makes is easy to join, extend, combine datasets and hence work collectively, but

  2. Towards interoperable and reproducible QSAR analyses: Exchange of datasets

    Directory of Open Access Journals (Sweden)

    Spjuth Ola

    2010-06-01

    Full Text Available Abstract Background QSAR is a widely used method to relate chemical structures to responses or properties based on experimental observations. Much effort has been made to evaluate and validate the statistical modeling in QSAR, but these analyses treat the dataset as fixed. An overlooked but highly important issue is the validation of the setup of the dataset, which comprises addition of chemical structures as well as selection of descriptors and software implementations prior to calculations. This process is hampered by the lack of standards and exchange formats in the field, making it virtually impossible to reproduce and validate analyses and drastically constrain collaborations and re-use of data. Results We present a step towards standardizing QSAR analyses by defining interoperable and reproducible QSAR datasets, consisting of an open XML format (QSAR-ML which builds on an open and extensible descriptor ontology. The ontology provides an extensible way of uniquely defining descriptors for use in QSAR experiments, and the exchange format supports multiple versioned implementations of these descriptors. Hence, a dataset described by QSAR-ML makes its setup completely reproducible. We also provide a reference implementation as a set of plugins for Bioclipse which simplifies setup of QSAR datasets, and allows for exporting in QSAR-ML as well as old-fashioned CSV formats. The implementation facilitates addition of new descriptor implementations from locally installed software and remote Web services; the latter is demonstrated with REST and XMPP Web services. Conclusions Standardized QSAR datasets open up new ways to store, query, and exchange data for subsequent analyses. QSAR-ML supports completely reproducible creation of datasets, solving the problems of defining which software components were used and their versions, and the descriptor ontology eliminates confusions regarding descriptors by defining them crisply. This makes is easy to join

  3. The Wind Integration National Dataset (WIND) toolkit (Presentation)

    Energy Technology Data Exchange (ETDEWEB)

    Caroline Draxl: NREL

    2014-01-01

    Regional wind integration studies require detailed wind power output data at many locations to perform simulations of how the power system will operate under high penetration scenarios. The wind datasets that serve as inputs into the study must realistically reflect the ramping characteristics, spatial and temporal correlations, and capacity factors of the simulated wind plants, as well as being time synchronized with available load profiles.As described in this presentation, the WIND Toolkit fulfills these requirements by providing a state-of-the-art national (US) wind resource, power production and forecast dataset.

  4. The Chandra Source Catalog 2.0: Estimating Source Fluxes

    Science.gov (United States)

    Primini, Francis Anthony; Allen, Christopher E.; Miller, Joseph; Anderson, Craig S.; Budynkiewicz, Jamie A.; Burke, Douglas; Chen, Judy C.; Civano, Francesca Maria; D'Abrusco, Raffaele; Doe, Stephen M.; Evans, Ian N.; Evans, Janet D.; Fabbiano, Giuseppina; Gibbs, Danny G., II; Glotfelty, Kenny J.; Graessle, Dale E.; Grier, John D.; Hain, Roger; Hall, Diane M.; Harbo, Peter N.; Houck, John C.; Lauer, Jennifer L.; Laurino, Omar; Lee, Nicholas P.; Martínez-Galarza, Juan Rafael; McCollough, Michael L.; McDowell, Jonathan C.; McLaughlin, Warren; Morgan, Douglas L.; Mossman, Amy E.; Nguyen, Dan T.; Nichols, Joy S.; Nowak, Michael A.; Paxson, Charles; Plummer, David A.; Rots, Arnold H.; Siemiginowska, Aneta; Sundheim, Beth A.; Tibbetts, Michael; Van Stone, David W.; Zografou, Panagoula

    2018-01-01

    The Second Chandra Source Catalog (CSC2.0) will provide information on approximately 316,000 point or compact extended x-ray sources, derived from over 10,000 ACIS and HRC-I imaging observations available in the public archive at the end of 2014. As in the previous catalog release (CSC1.1), fluxes for these sources will be determined separately from source detection, using a Bayesian formalism that accounts for background, spatial resolution effects, and contamination from nearby sources. However, the CSC2.0 procedure differs from that used in CSC1.1 in three important aspects. First, for sources in crowded regions in which photometric apertures overlap, fluxes are determined jointly, using an extension of the CSC1.1 algorithm, as discussed in Primini & Kashyap (2014ApJ...796…24P). Second, an MCMC procedure is used to estimate marginalized posterior probability distributions for source fluxes. Finally, for sources observed in multiple observations, a Bayesian Blocks algorithm (Scargle, et al. 2013ApJ...764..167S) is used to group observations into blocks of constant source flux.In this poster we present details of the CSC2.0 photometry algorithms and illustrate their performance in actual CSC2.0 datasets.This work has been supported by NASA under contract NAS 8-03060 to the Smithsonian Astrophysical Observatory for operation of the Chandra X-ray Center.

  5. Monte Carlo surface flux tallies

    International Nuclear Information System (INIS)

    Favorite, Jeffrey A.

    2010-01-01

    Particle fluxes on surfaces are difficult to calculate with Monte Carlo codes because the score requires a division by the surface-crossing angle cosine, and grazing angles lead to inaccuracies. We revisit the standard practice of dividing by half of a cosine 'cutoff' for particles whose surface-crossing cosines are below the cutoff. The theory behind this approximation is sound, but the application of the theory to all possible situations does not account for two implicit assumptions: (1) the grazing band must be symmetric about 0, and (2) a single linear expansion for the angular flux must be applied in the entire grazing band. These assumptions are violated in common circumstances; for example, for separate in-going and out-going flux tallies on internal surfaces, and for out-going flux tallies on external surfaces. In some situations, dividing by two-thirds of the cosine cutoff is more appropriate. If users were able to control both the cosine cutoff and the substitute value, they could use these parameters to make accurate surface flux tallies. The procedure is demonstrated in a test problem in which Monte Carlo surface fluxes in cosine bins are converted to angular fluxes and compared with the results of a discrete ordinates calculation.

  6. Using Multiple Big Datasets and Machine Learning to Produce a New Global Particulate Dataset: A Technology Challenge Case Study

    Science.gov (United States)

    Lary, D. J.

    2013-12-01

    A BigData case study is described where multiple datasets from several satellites, high-resolution global meteorological data, social media and in-situ observations are combined using machine learning on a distributed cluster using an automated workflow. The global particulate dataset is relevant to global public health studies and would not be possible to produce without the use of the multiple big datasets, in-situ data and machine learning.To greatly reduce the development time and enhance the functionality a high level language capable of parallel processing has been used (Matlab). A key consideration for the system is high speed access due to the large data volume, persistence of the large data volumes and a precise process time scheduling capability.

  7. Would the ‘real’ observed dataset stand up? A critical examination of eight observed gridded climate datasets for China

    International Nuclear Information System (INIS)

    Sun, Qiaohong; Miao, Chiyuan; Duan, Qingyun; Kong, Dongxian; Ye, Aizhong; Di, Zhenhua; Gong, Wei

    2014-01-01

    This research compared and evaluated the spatio-temporal similarities and differences of eight widely used gridded datasets. The datasets include daily precipitation over East Asia (EA), the Climate Research Unit (CRU) product, the Global Precipitation Climatology Centre (GPCC) product, the University of Delaware (UDEL) product, Precipitation Reconstruction over Land (PREC/L), the Asian Precipitation Highly Resolved Observational (APHRO) product, the Institute of Atmospheric Physics (IAP) dataset from the Chinese Academy of Sciences, and the National Meteorological Information Center dataset from the China Meteorological Administration (CN05). The meteorological variables focus on surface air temperature (SAT) or precipitation (PR) in China. All datasets presented general agreement on the whole spatio-temporal scale, but some differences appeared for specific periods and regions. On a temporal scale, EA shows the highest amount of PR, while APHRO shows the lowest. CRU and UDEL show higher SAT than IAP or CN05. On a spatial scale, the most significant differences occur in western China for PR and SAT. For PR, the difference between EA and CRU is the largest. When compared with CN05, CRU shows higher SAT in the central and southern Northwest river drainage basin, UDEL exhibits higher SAT over the Southwest river drainage system, and IAP has lower SAT in the Tibetan Plateau. The differences in annual mean PR and SAT primarily come from summer and winter, respectively. Finally, potential factors impacting agreement among gridded climate datasets are discussed, including raw data sources, quality control (QC) schemes, orographic correction, and interpolation techniques. The implications and challenges of these results for climate research are also briefly addressed. (paper)

  8. Measurements of the Canonical Helicity Evolution of a Gyrating Kinked Flux Rope

    Science.gov (United States)

    von der Linden, J.; Sears, J.; Intrator, T.; You, S.

    2017-12-01

    Magnetic structures in the solar corona and planetary magnetospheres are often modelled as magnetic flux ropes governed by magnetohydrodynamics (MHD); however, inside these structures, as exhibited in reconnection, conversions between magnetic and kinetic energies occur over a wide range of scales. Flux ropes based on the flux of canonical momentum circulation extend the flux rope concept to include effects of finite particle momentum and present the distinct advantage of reconciling all plasma regimes - e.g. kinetic, two-fluid, and MHD - with the topological concept of helicity: twists, writhes, and linkages. This presentation shows the first visualization and analysis of the 3D dynamics of canonical flux ropes and their relative helicity evolution from laboratory measurements. Ion and electron canonical flux ropes are visualized from a dataset of Mach, triple, and Ḃ probe measurements at over 10,000 spatial locations of a gyrating kinked flux rope. The flux ropes co-gyrate with the peak density and electron temperature in and out of a measurement volume. The electron and ion canonical flux ropes twist with opposite handedness and the ion flux ropes writhe around the electron flux ropes. The relative cross helicity between the magnetic and ion flow vorticity flux ropes dominates the relative ion canonical helicity and is anti-correlated with the relative magnetic helicity. The 3D nature of the kink and a reverse eddy current affect the helicity evolution. This work is supported by DOE Grant DE-SC0010340 and the DOE Office of Science Graduate Student Research Program and prepared in part by LLNL under Contract DE-AC52-07NA27344. LLNL-ABS-735426

  9. Using Real Datasets for Interdisciplinary Business/Economics Projects

    Science.gov (United States)

    Goel, Rajni; Straight, Ronald L.

    2005-01-01

    The workplace's global and dynamic nature allows and requires improved approaches for providing business and economics education. In this article, the authors explore ways of enhancing students' understanding of course material by using nontraditional, real-world datasets of particular interest to them. Teaching at a historically Black university,…

  10. Dataset-driven research for improving recommender systems for learning

    NARCIS (Netherlands)

    Verbert, Katrien; Drachsler, Hendrik; Manouselis, Nikos; Wolpers, Martin; Vuorikari, Riina; Duval, Erik

    2011-01-01

    Verbert, K., Drachsler, H., Manouselis, N., Wolpers, M., Vuorikari, R., & Duval, E. (2011). Dataset-driven research for improving recommender systems for learning. In Ph. Long, & G. Siemens (Eds.), Proceedings of 1st International Conference Learning Analytics & Knowledge (pp. 44-53). February,

  11. dataTEL - Datasets for Technology Enhanced Learning

    NARCIS (Netherlands)

    Drachsler, Hendrik; Verbert, Katrien; Sicilia, Miguel-Angel; Wolpers, Martin; Manouselis, Nikos; Vuorikari, Riina; Lindstaedt, Stefanie; Fischer, Frank

    2011-01-01

    Drachsler, H., Verbert, K., Sicilia, M. A., Wolpers, M., Manouselis, N., Vuorikari, R., Lindstaedt, S., & Fischer, F. (2011). dataTEL - Datasets for Technology Enhanced Learning. STELLAR Alpine Rendez-Vous White Paper. Alpine Rendez-Vous 2011 White paper collection, Nr. 13., France (2011)

  12. A dataset of forest biomass structure for Eurasia.

    Science.gov (United States)

    Schepaschenko, Dmitry; Shvidenko, Anatoly; Usoltsev, Vladimir; Lakyda, Petro; Luo, Yunjian; Vasylyshyn, Roman; Lakyda, Ivan; Myklush, Yuriy; See, Linda; McCallum, Ian; Fritz, Steffen; Kraxner, Florian; Obersteiner, Michael

    2017-05-16

    The most comprehensive dataset of in situ destructive sampling measurements of forest biomass in Eurasia have been compiled from a combination of experiments undertaken by the authors and from scientific publications. Biomass is reported as four components: live trees (stem, bark, branches, foliage, roots); understory (above- and below ground); green forest floor (above- and below ground); and coarse woody debris (snags, logs, dead branches of living trees and dead roots), consisting of 10,351 unique records of sample plots and 9,613 sample trees from ca 1,200 experiments for the period 1930-2014 where there is overlap between these two datasets. The dataset also contains other forest stand parameters such as tree species composition, average age, tree height, growing stock volume, etc., when available. Such a dataset can be used for the development of models of biomass structure, biomass extension factors, change detection in biomass structure, investigations into biodiversity and species distribution and the biodiversity-productivity relationship, as well as the assessment of the carbon pool and its dynamics, among many others.

  13. A reanalysis dataset of the South China Sea

    Science.gov (United States)

    Zeng, Xuezhi; Peng, Shiqiu; Li, Zhijin; Qi, Yiquan; Chen, Rongyu

    2014-01-01

    Ocean reanalysis provides a temporally continuous and spatially gridded four-dimensional estimate of the ocean state for a better understanding of the ocean dynamics and its spatial/temporal variability. Here we present a 19-year (1992–2010) high-resolution ocean reanalysis dataset of the upper ocean in the South China Sea (SCS) produced from an ocean data assimilation system. A wide variety of observations, including in-situ temperature/salinity profiles, ship-measured and satellite-derived sea surface temperatures, and sea surface height anomalies from satellite altimetry, are assimilated into the outputs of an ocean general circulation model using a multi-scale incremental three-dimensional variational data assimilation scheme, yielding a daily high-resolution reanalysis dataset of the SCS. Comparisons between the reanalysis and independent observations support the reliability of the dataset. The presented dataset provides the research community of the SCS an important data source for studying the thermodynamic processes of the ocean circulation and meso-scale features in the SCS, including their spatial and temporal variability. PMID:25977803

  14. Comparision of analysis of the QTLMAS XII common dataset

    DEFF Research Database (Denmark)

    Crooks, Lucy; Sahana, Goutam; de Koning, Dirk-Jan

    2009-01-01

    As part of the QTLMAS XII workshop, a simulated dataset was distributed and participants were invited to submit analyses of the data based on genome-wide association, fine mapping and genomic selection. We have evaluated the findings from the groups that reported fine mapping and genome-wide asso...

  15. The LAMBADA dataset: Word prediction requiring a broad discourse context

    NARCIS (Netherlands)

    Paperno, D.; Kruszewski, G.; Lazaridou, A.; Pham, Q.N.; Bernardi, R.; Pezzelle, S.; Baroni, M.; Boleda, G.; Fernández, R.; Erk, K.; Smith, N.A.

    2016-01-01

    We introduce LAMBADA, a dataset to evaluate the capabilities of computational models for text understanding by means of a word prediction task. LAMBADA is a collection of narrative passages sharing the characteristic that human subjects are able to guess their last word if they are exposed to the

  16. NEW WEB-BASED ACCESS TO NUCLEAR STRUCTURE DATASETS.

    Energy Technology Data Exchange (ETDEWEB)

    WINCHELL,D.F.

    2004-09-26

    As part of an effort to migrate the National Nuclear Data Center (NNDC) databases to a relational platform, a new web interface has been developed for the dissemination of the nuclear structure datasets stored in the Evaluated Nuclear Structure Data File and Experimental Unevaluated Nuclear Data List.

  17. Cross-Cultural Concept Mapping of Standardized Datasets

    DEFF Research Database (Denmark)

    Kano Glückstad, Fumiko

    2012-01-01

    This work compares four feature-based similarity measures derived from cognitive sciences. The purpose of the comparative analysis is to verify the potentially most effective model that can be applied for mapping independent ontologies in a culturally influenced domain [1]. Here, datasets based...

  18. Level-1 muon trigger performance with the full 2017 dataset

    CERN Document Server

    CMS Collaboration

    2018-01-01

    This document describes the performance of the CMS Level-1 Muon Trigger with the full dataset of 2017. Efficiency plots are included for each track finder (TF) individually and for the system as a whole. The efficiency is measured to be greater than 90% for all track finders.

  19. A Dataset for Visual Navigation with Neuromorphic Methods

    Directory of Open Access Journals (Sweden)

    Francisco eBarranco

    2016-02-01

    Full Text Available Standardized benchmarks in Computer Vision have greatly contributed to the advance of approaches to many problems in the field. If we want to enhance the visibility of event-driven vision and increase its impact, we will need benchmarks that allow comparison among different neuromorphic methods as well as comparison to Computer Vision conventional approaches. We present datasets to evaluate the accuracy of frame-free and frame-based approaches for tasks of visual navigation. Similar to conventional Computer Vision datasets, we provide synthetic and real scenes, with the synthetic data created with graphics packages, and the real data recorded using a mobile robotic platform carrying a dynamic and active pixel vision sensor (DAVIS and an RGB+Depth sensor. For both datasets the cameras move with a rigid motion in a static scene, and the data includes the images, events, optic flow, 3D camera motion, and the depth of the scene, along with calibration procedures. Finally, we also provide simulated event data generated synthetically from well-known frame-based optical flow datasets.

  20. Evaluation of Uncertainty in Precipitation Datasets for New Mexico, USA

    Science.gov (United States)

    Besha, A. A.; Steele, C. M.; Fernald, A.

    2014-12-01

    Climate change, population growth and other factors are endangering water availability and sustainability in semiarid/arid areas particularly in the southwestern United States. Wide coverage of spatial and temporal measurements of precipitation are key for regional water budget analysis and hydrological operations which themselves are valuable tool for water resource planning and management. Rain gauge measurements are usually reliable and accurate at a point. They measure rainfall continuously, but spatial sampling is limited. Ground based radar and satellite remotely sensed precipitation have wide spatial and temporal coverage. However, these measurements are indirect and subject to errors because of equipment, meteorological variability, the heterogeneity of the land surface itself and lack of regular recording. This study seeks to understand precipitation uncertainty and in doing so, lessen uncertainty propagation into hydrological applications and operations. We reviewed, compared and evaluated the TRMM (Tropical Rainfall Measuring Mission) precipitation products, NOAA's (National Oceanic and Atmospheric Administration) Global Precipitation Climatology Centre (GPCC) monthly precipitation dataset, PRISM (Parameter elevation Regression on Independent Slopes Model) data and data from individual climate stations including Cooperative Observer Program (COOP), Remote Automated Weather Stations (RAWS), Soil Climate Analysis Network (SCAN) and Snowpack Telemetry (SNOTEL) stations. Though not yet finalized, this study finds that the uncertainty within precipitation estimates datasets is influenced by regional topography, season, climate and precipitation rate. Ongoing work aims to further evaluate precipitation datasets based on the relative influence of these phenomena so that we can identify the optimum datasets for input to statewide water budget analysis.

  1. Dataset: Multi Sensor-Orientation Movement Data of Goats

    NARCIS (Netherlands)

    Kamminga, Jacob Wilhelm

    2018-01-01

    This is a labeled dataset. Motion data were collected from six sensor nodes that were fixed with different orientations to a collar around the neck of goats. These six sensor nodes simultaneously, with different orientations, recorded various activities performed by the goat. We recorded the

  2. A dataset of human decision-making in teamwork management

    Science.gov (United States)

    Yu, Han; Shen, Zhiqi; Miao, Chunyan; Leung, Cyril; Chen, Yiqiang; Fauvel, Simon; Lin, Jun; Cui, Lizhen; Pan, Zhengxiang; Yang, Qiang

    2017-01-01

    Today, most endeavours require teamwork by people with diverse skills and characteristics. In managing teamwork, decisions are often made under uncertainty and resource constraints. The strategies and the effectiveness of the strategies different people adopt to manage teamwork under different situations have not yet been fully explored, partially due to a lack of detailed large-scale data. In this paper, we describe a multi-faceted large-scale dataset to bridge this gap. It is derived from a game simulating complex project management processes. It presents the participants with different conditions in terms of team members' capabilities and task characteristics for them to exhibit their decision-making strategies. The dataset contains detailed data reflecting the decision situations, decision strategies, decision outcomes, and the emotional responses of 1,144 participants from diverse backgrounds. To our knowledge, this is the first dataset simultaneously covering these four facets of decision-making. With repeated measurements, the dataset may help establish baseline variability of decision-making in teamwork management, leading to more realistic decision theoretic models and more effective decision support approaches.

  3. UK surveillance: provision of quality assured information from combined datasets.

    Science.gov (United States)

    Paiba, G A; Roberts, S R; Houston, C W; Williams, E C; Smith, L H; Gibbens, J C; Holdship, S; Lysons, R

    2007-09-14

    Surveillance information is most useful when provided within a risk framework, which is achieved by presenting results against an appropriate denominator. Often the datasets are captured separately and for different purposes, and will have inherent errors and biases that can be further confounded by the act of merging. The United Kingdom Rapid Analysis and Detection of Animal-related Risks (RADAR) system contains data from several sources and provides both data extracts for research purposes and reports for wider stakeholders. Considerable efforts are made to optimise the data in RADAR during the Extraction, Transformation and Loading (ETL) process. Despite efforts to ensure data quality, the final dataset inevitably contains some data errors and biases, most of which cannot be rectified during subsequent analysis. So, in order for users to establish the 'fitness for purpose' of data merged from more than one data source, Quality Statements are produced as defined within the overarching surveillance Quality Framework. These documents detail identified data errors and biases following ETL and report construction as well as relevant aspects of the datasets from which the data originated. This paper illustrates these issues using RADAR datasets, and describes how they can be minimised.

  4. participatory development of a minimum dataset for the khayelitsha ...

    African Journals Online (AJOL)

    This dataset was integrated with data requirements at ... model for defining health information needs at district level. This participatory process has enabled health workers to appraise their .... of reproductive health, mental health, disability and community ... each chose a facilitator and met in between the forum meetings.

  5. Comparision of analysis of the QTLMAS XII common dataset

    DEFF Research Database (Denmark)

    Lund, Mogens Sandø; Sahana, Goutam; de Koning, Dirk-Jan

    2009-01-01

    A dataset was simulated and distributed to participants of the QTLMAS XII workshop who were invited to develop genomic selection models. Each contributing group was asked to describe the model development and validation as well as to submit genomic predictions for three generations of individuals...

  6. The NASA Subsonic Jet Particle Image Velocimetry (PIV) Dataset

    Science.gov (United States)

    Bridges, James; Wernet, Mark P.

    2011-01-01

    Many tasks in fluids engineering require prediction of turbulence of jet flows. The present document documents the single-point statistics of velocity, mean and variance, of cold and hot jet flows. The jet velocities ranged from 0.5 to 1.4 times the ambient speed of sound, and temperatures ranged from unheated to static temperature ratio 2.7. Further, the report assesses the accuracies of the data, e.g., establish uncertainties for the data. This paper covers the following five tasks: (1) Document acquisition and processing procedures used to create the particle image velocimetry (PIV) datasets. (2) Compare PIV data with hotwire and laser Doppler velocimetry (LDV) data published in the open literature. (3) Compare different datasets acquired at the same flow conditions in multiple tests to establish uncertainties. (4) Create a consensus dataset for a range of hot jet flows, including uncertainty bands. (5) Analyze this consensus dataset for self-consistency and compare jet characteristics to those of the open literature. The final objective was fulfilled by using the potential core length and the spread rate of the half-velocity radius to collapse of the mean and turbulent velocity fields over the first 20 jet diameters.

  7. A new dataset validation system for the Planetary Science Archive

    Science.gov (United States)

    Manaud, N.; Zender, J.; Heather, D.; Martinez, S.

    2007-08-01

    The Planetary Science Archive is the official archive for the Mars Express mission. It has received its first data by the end of 2004. These data are delivered by the PI teams to the PSA team as datasets, which are formatted conform to the Planetary Data System (PDS). The PI teams are responsible for analyzing and calibrating the instrument data as well as the production of reduced and calibrated data. They are also responsible of the scientific validation of these data. ESA is responsible of the long-term data archiving and distribution to the scientific community and must ensure, in this regard, that all archived products meet quality. To do so, an archive peer-review is used to control the quality of the Mars Express science data archiving process. However a full validation of its content is missing. An independent review board recently recommended that the completeness of the archive as well as the consistency of the delivered data should be validated following well-defined procedures. A new validation software tool is being developed to complete the overall data quality control system functionality. This new tool aims to improve the quality of data and services provided to the scientific community through the PSA, and shall allow to track anomalies in and to control the completeness of datasets. It shall ensure that the PSA end-users: (1) can rely on the result of their queries, (2) will get data products that are suitable for scientific analysis, (3) can find all science data acquired during a mission. We defined dataset validation as the verification and assessment process to check the dataset content against pre-defined top-level criteria, which represent the general characteristics of good quality datasets. The dataset content that is checked includes the data and all types of information that are essential in the process of deriving scientific results and those interfacing with the PSA database. The validation software tool is a multi-mission tool that

  8. Data Recommender: An Alternative Way to Discover Open Scientific Datasets

    Science.gov (United States)

    Klump, J. F.; Devaraju, A.; Williams, G.; Hogan, D.; Davy, R.; Page, J.; Singh, D.; Peterson, N.

    2017-12-01

    Over the past few years, institutions and government agencies have adopted policies to openly release their data, which has resulted in huge amounts of open data becoming available on the web. When trying to discover the data, users face two challenges: an overload of choice and the limitations of the existing data search tools. On the one hand, there are too many datasets to choose from, and therefore, users need to spend considerable effort to find the datasets most relevant to their research. On the other hand, data portals commonly offer keyword and faceted search, which depend fully on the user queries to search and rank relevant datasets. Consequently, keyword and faceted search may return loosely related or irrelevant results, although the results may contain the same query. They may also return highly specific results that depend more on how well metadata was authored. They do not account well for variance in metadata due to variance in author styles and preferences. The top-ranked results may also come from the same data collection, and users are unlikely to discover new and interesting datasets. These search modes mainly suits users who can express their information needs in terms of the structure and terminology of the data portals, but may pose a challenge otherwise. The above challenges reflect that we need a solution that delivers the most relevant (i.e., similar and serendipitous) datasets to users, beyond the existing search functionalities on the portals. A recommender system is an information filtering system that presents users with relevant and interesting contents based on users' context and preferences. Delivering data recommendations to users can make data discovery easier, and as a result may enhance user engagement with the portal. We developed a hybrid data recommendation approach for the CSIRO Data Access Portal. The approach leverages existing recommendation techniques (e.g., content-based filtering and item co-occurrence) to produce

  9. Comparison of global 3-D aviation emissions datasets

    Directory of Open Access Journals (Sweden)

    S. C. Olsen

    2013-01-01

    Full Text Available Aviation emissions are unique from other transportation emissions, e.g., from road transportation and shipping, in that they occur at higher altitudes as well as at the surface. Aviation emissions of carbon dioxide, soot, and water vapor have direct radiative impacts on the Earth's climate system while emissions of nitrogen oxides (NOx, sulfur oxides, carbon monoxide (CO, and hydrocarbons (HC impact air quality and climate through their effects on ozone, methane, and clouds. The most accurate estimates of the impact of aviation on air quality and climate utilize three-dimensional chemistry-climate models and gridded four dimensional (space and time aviation emissions datasets. We compare five available aviation emissions datasets currently and historically used to evaluate the impact of aviation on climate and air quality: NASA-Boeing 1992, NASA-Boeing 1999, QUANTIFY 2000, Aero2k 2002, and AEDT 2006 and aviation fuel usage estimates from the International Energy Agency. Roughly 90% of all aviation emissions are in the Northern Hemisphere and nearly 60% of all fuelburn and NOx emissions occur at cruise altitudes in the Northern Hemisphere. While these datasets were created by independent methods and are thus not strictly suitable for analyzing trends they suggest that commercial aviation fuelburn and NOx emissions increased over the last two decades while HC emissions likely decreased and CO emissions did not change significantly. The bottom-up estimates compared here are consistently lower than International Energy Agency fuelburn statistics although the gap is significantly smaller in the more recent datasets. Overall the emissions distributions are quite similar for fuelburn and NOx with regional peaks over the populated land masses of North America, Europe, and East Asia. For CO and HC there are relatively larger differences. There are however some distinct differences in the altitude distribution

  10. Geoseq: a tool for dissecting deep-sequencing datasets

    Directory of Open Access Journals (Sweden)

    Homann Robert

    2010-10-01

    Full Text Available Abstract Background Datasets generated on deep-sequencing platforms have been deposited in various public repositories such as the Gene Expression Omnibus (GEO, Sequence Read Archive (SRA hosted by the NCBI, or the DNA Data Bank of Japan (ddbj. Despite being rich data sources, they have not been used much due to the difficulty in locating and analyzing datasets of interest. Results Geoseq http://geoseq.mssm.edu provides a new method of analyzing short reads from deep sequencing experiments. Instead of mapping the reads to reference genomes or sequences, Geoseq maps a reference sequence against the sequencing data. It is web-based, and holds pre-computed data from public libraries. The analysis reduces the input sequence to tiles and measures the coverage of each tile in a sequence library through the use of suffix arrays. The user can upload custom target sequences or use gene/miRNA names for the search and get back results as plots and spreadsheet files. Geoseq organizes the public sequencing data using a controlled vocabulary, allowing identification of relevant libraries by organism, tissue and type of experiment. Conclusions Analysis of small sets of sequences against deep-sequencing datasets, as well as identification of public datasets of interest, is simplified by Geoseq. We applied Geoseq to, a identify differential isoform expression in mRNA-seq datasets, b identify miRNAs (microRNAs in libraries, and identify mature and star sequences in miRNAS and c to identify potentially mis-annotated miRNAs. The ease of using Geoseq for these analyses suggests its utility and uniqueness as an analysis tool.

  11. Conical electromagnetic radiation flux concentrator

    Science.gov (United States)

    Miller, E. R.

    1972-01-01

    Concentrator provides method of concentrating a beam of electromagnetic radiation into a smaller beam, presenting a higher flux density. Smaller beam may be made larger by sending radiation through the device in the reverse direction.

  12. Physics of Magnetic Flux Ropes

    CERN Document Server

    Priest, E R; Lee, L C

    1990-01-01

    The American Geophysical Union Chapman Conference on the Physics of Magnetic Flux Ropes was held at the Hamilton Princess Hotel, Hamilton, Bermuda on March 27–31, 1989. Topics discussed ranged from solar flux ropes, such as photospheric flux tubes, coronal loops and prominences, to flux ropes in the solar wind, in planetary ionospheres, at the Earth's magnetopause, in the geomagnetic tail and deep in the Earth's magnetosphere. Papers presented at that conference form the nucleus of this book, but the book is more than just a proceedings of the conference. We have solicited articles from all interested in this topic. Thus, there is some material in the book not discussed at the conference. Even in the case of papers presented at the conference, there is generally a much more detailed and rigorous presentation than was possible in the time allowed by the oral and poster presentations.

  13. Notes on neutron flux measurement

    International Nuclear Information System (INIS)

    Alcala Ruiz, F.

    1984-01-01

    The main purpose of this work is to get an useful guide to carry out topical neutron flux measurements. Although the foil activation technique is used in the majority of the cases, other techniques, such as those based on fission chambers and self-powered neutron detectors, are also shown. Special interest is given to the description and application of corrections on the measurement of relative and absolute induced activities by several types of detectors (scintillators, G-M and gas proportional counters). The thermal arid epithermal neutron fluxes, as determined in this work, are conventional or effective (West cots fluxes), which are extensively used by the reactor experimentalists; however, we also give some expressions where they are related to the integrated neutron fluxes, which are used in neutron calculations. (Author) 16 refs

  14. Specification of ROP flux shape

    International Nuclear Information System (INIS)

    Min, Byung Joo; Gray, A.

    1997-06-01

    The CANDU 9 480/SEU core uses 0.9% SEU (Slightly Enriched Uranium) fuel. The use f SEU fuel enables the reactor to increase the radial power form factor from 0.865, which is typical in current natural uranium CANDU reactors, to 0.97 in the nominal CANDU 9 480/SEU core. The difference is a 12% increase in reactor power. An additional 5% increase can be achieved due to a reduced refuelling ripple. The channel power limits were also increased by 3% for a total reactor power increase of 20%. This report describes the calculation of neutron flux distributions in the CANDU 9 480/SEU core under conditions specified by the C and I engineers. The RFSP code was used to calculate of neutron flux shapes for ROP analysis. Detailed flux values at numerous potential detector sites were calculated for each flux shape. (author). 6 tabs., 70 figs., 4 refs

  15. Specification of ROP flux shape

    Energy Technology Data Exchange (ETDEWEB)

    Min, Byung Joo [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of); Gray, A [Atomic Energy of Canada Ltd., Chalk River, ON (Canada)

    1997-06-01

    The CANDU 9 480/SEU core uses 0.9% SEU (Slightly Enriched Uranium) fuel. The use f SEU fuel enables the reactor to increase the radial power form factor from 0.865, which is typical in current natural uranium CANDU reactors, to 0.97 in the nominal CANDU 9 480/SEU core. The difference is a 12% increase in reactor power. An additional 5% increase can be achieved due to a reduced refuelling ripple. The channel power limits were also increased by 3% for a total reactor power increase of 20%. This report describes the calculation of neutron flux distributions in the CANDU 9 480/SEU core under conditions specified by the C and I engineers. The RFSP code was used to calculate of neutron flux shapes for ROP analysis. Detailed flux values at numerous potential detector sites were calculated for each flux shape. (author). 6 tabs., 70 figs., 4 refs.

  16. High Flux Isotope Reactor (HFIR)

    Data.gov (United States)

    Federal Laboratory Consortium — The HFIR at Oak Ridge National Laboratory is a light-water cooled and moderated reactor that is the United States’ highest flux reactor-based neutron source. HFIR...

  17. Flux networks in metabolic graphs

    International Nuclear Information System (INIS)

    Warren, P B; Queiros, S M Duarte; Jones, J L

    2009-01-01

    A metabolic model can be represented as a bipartite graph comprising linked reaction and metabolite nodes. Here it is shown how a network of conserved fluxes can be assigned to the edges of such a graph by combining the reaction fluxes with a conserved metabolite property such as molecular weight. A similar flux network can be constructed by combining the primal and dual solutions to the linear programming problem that typically arises in constraint-based modelling. Such constructions may help with the visualization of flux distributions in complex metabolic networks. The analysis also explains the strong correlation observed between metabolite shadow prices (the dual linear programming variables) and conserved metabolite properties. The methods were applied to recent metabolic models for Escherichia coli, Saccharomyces cerevisiae and Methanosarcina barkeri. Detailed results are reported for E. coli; similar results were found for other organisms

  18. On sample size and different interpretations of snow stability datasets

    Science.gov (United States)

    Schirmer, M.; Mitterer, C.; Schweizer, J.

    2009-04-01

    Interpretations of snow stability variations need an assessment of the stability itself, independent of the scale investigated in the study. Studies on stability variations at a regional scale have often chosen stability tests such as the Rutschblock test or combinations of various tests in order to detect differences in aspect and elevation. The question arose: ‘how capable are such stability interpretations in drawing conclusions'. There are at least three possible errors sources: (i) the variance of the stability test itself; (ii) the stability variance at an underlying slope scale, and (iii) that the stability interpretation might not be directly related to the probability of skier triggering. Various stability interpretations have been proposed in the past that provide partly different results. We compared a subjective one based on expert knowledge with a more objective one based on a measure derived from comparing skier-triggered slopes vs. slopes that have been skied but not triggered. In this study, the uncertainties are discussed and their effects on regional scale stability variations will be quantified in a pragmatic way. An existing dataset with very large sample sizes was revisited. This dataset contained the variance of stability at a regional scale for several situations. The stability in this dataset was determined using the subjective interpretation scheme based on expert knowledge. The question to be answered was how many measurements were needed to obtain similar results (mainly stability differences in aspect or elevation) as with the complete dataset. The optimal sample size was obtained in several ways: (i) assuming a nominal data scale the sample size was determined with a given test, significance level and power, and by calculating the mean and standard deviation of the complete dataset. With this method it can also be determined if the complete dataset consists of an appropriate sample size. (ii) Smaller subsets were created with similar

  19. Boundary fluxes for nonlocal diffusion

    Science.gov (United States)

    Cortazar, Carmen; Elgueta, Manuel; Rossi, Julio D.; Wolanski, Noemi

    We study a nonlocal diffusion operator in a bounded smooth domain prescribing the flux through the boundary. This problem may be seen as a generalization of the usual Neumann problem for the heat equation. First, we prove existence, uniqueness and a comparison principle. Next, we study the behavior of solutions for some prescribed boundary data including blowing up ones. Finally, we look at a nonlinear flux boundary condition.

  20. Energy flux of hot atoms

    International Nuclear Information System (INIS)

    Wotzak, G.P.; Kostin, M.D.

    1976-01-01

    The process in which hot atoms collide with thermal atoms of a gas, transfer kinetic energy to them, and produce additional hot atoms is investigated. A stochastic method is used to obtain numerical results for the spatial and time dependent energy flux of hot atoms in a gas. The results indicate that in hot atom systems a front followed by an intense energy flux of hot atoms may develop

  1. Flux tubes at finite temperature

    Energy Technology Data Exchange (ETDEWEB)

    Cea, Paolo [INFN, Sezione di Bari,Via G. Amendola 173, I-70126 Bari (Italy); Dipartimento di Fisica dell’Università di Bari,Via G. Amendola 173, I-70126 Bari (Italy); Cosmai, Leonardo [INFN, Sezione di Bari,Via G. Amendola 173, I-70126 Bari (Italy); Cuteri, Francesca; Papa, Alessandro [Dipartimento di Fisica, Università della Calabria & INFN-Cosenza,Ponte Bucci, cubo 31C, I-87036 Rende (Cosenza) (Italy)

    2016-06-07

    The chromoelectric field generated by a static quark-antiquark pair, with its peculiar tube-like shape, can be nicely described, at zero temperature, within the dual superconductor scenario for the QCD confining vacuum. In this work we investigate, by lattice Monte Carlo simulations of the SU(3) pure gauge theory, the fate of chromoelectric flux tubes across the deconfinement transition. We find that, if the distance between the static sources is kept fixed at about 0.76 fm ≃1.6/√σ and the temperature is increased towards and above the deconfinement temperature T{sub c}, the amplitude of the field inside the flux tube gets smaller, while the shape of the flux tube does not vary appreciably across deconfinement. This scenario with flux-tube “evaporation” above T{sub c} has no correspondence in ordinary (type-II) superconductivity, where instead the transition to the phase with normal conductivity is characterized by a divergent fattening of flux tubes as the transition temperature is approached from below. We present also some evidence about the existence of flux-tube structures in the magnetic sector of the theory in the deconfined phase.

  2. P fluxes and exotic branes

    Energy Technology Data Exchange (ETDEWEB)

    Lombardo, Davide M. [Dipartimento di Fisica, Università di Roma “La Sapienza”,Piazzale Aldo Moro 2, 00185 Roma (Italy); Riccioni, Fabio [INFN - Sezione di Roma, Dipartimento di Fisica, Università di Roma “La Sapienza”,Piazzale Aldo Moro 2, 00185 Roma (Italy); Risoli, Stefano [Dipartimento di Fisica, Università di Roma “La Sapienza”,Piazzale Aldo Moro 2, 00185 Roma (Italy); INFN - Sezione di Roma, Dipartimento di Fisica, Università di Roma “La Sapienza”,Piazzale Aldo Moro 2, 00185 Roma (Italy)

    2016-12-21

    We consider the N=1 superpotential generated in type-II orientifold models by non-geometric fluxes. In particular, we focus on the family of P fluxes, that are related by T-duality transformations to the S-dual of the Q flux. We determine the general rule that transforms a given flux in this family under a single T-duality transformation. This rule allows to derive a complete expression for the superpotential for both the IIA and the IIB theory for the particular case of a T{sup 6}/[ℤ{sub 2}×ℤ{sub 2}] orientifold. We then consider how these fluxes modify the generalised Bianchi identities. In particular, we derive a fully consistent set of quadratic constraints coming from the NS-NS Bianchi identities. On the other hand, the P flux Bianchi identities induce tadpoles, and we determine a set of exotic branes that can be consistently included in order to cancel them. This is achieved by determining a universal transformation rule under T-duality satisfied by all the branes in string theory.

  3. P fluxes and exotic branes

    International Nuclear Information System (INIS)

    Lombardo, Davide M.; Riccioni, Fabio; Risoli, Stefano

    2016-01-01

    We consider the N=1 superpotential generated in type-II orientifold models by non-geometric fluxes. In particular, we focus on the family of P fluxes, that are related by T-duality transformations to the S-dual of the Q flux. We determine the general rule that transforms a given flux in this family under a single T-duality transformation. This rule allows to derive a complete expression for the superpotential for both the IIA and the IIB theory for the particular case of a T 6 /[ℤ 2 ×ℤ 2 ] orientifold. We then consider how these fluxes modify the generalised Bianchi identities. In particular, we derive a fully consistent set of quadratic constraints coming from the NS-NS Bianchi identities. On the other hand, the P flux Bianchi identities induce tadpoles, and we determine a set of exotic branes that can be consistently included in order to cancel them. This is achieved by determining a universal transformation rule under T-duality satisfied by all the branes in string theory.

  4. Establishment and analysis of a High-Resolution Assimilation Dataset of the water-energy cycle in China

    Science.gov (United States)

    Zhu, X.; Wen, X.; Zheng, Z.

    2017-12-01

    For better prediction and understanding of land-atmospheric interaction, in-situ observed meteorological data acquired from the China Meteorological Administration (CMA) were assimilated in the Weather Research and Forecasting (WRF) model and the monthly Green Vegetation Coverage (GVF) data, which was calculated using the Normalized Difference Vegetation Index (NDVI) of the Earth Observing System Moderate-Resolution Imaging Spectroradiometer (EOS-MODIS) and Digital Elevation Model (DEM) data of the Shuttle Radar Topography Mission (SRTM) system. Furthermore, the WRF model produced a High-Resolution Assimilation Dataset of the water-energy cycle in China (HRADC). This dataset has a horizontal resolution of 25 km for near surface meteorological data, such as air temperature, humidity, wind vectors and pressure (19 levels); soil temperature and moisture (four levels); surface temperature; downward/upward short/long radiation; 3-h latent heat flux; sensible heat flux; and ground heat flux. In this study, we 1) briefly introduce the cycling 3D-Var assimilation method and 2) compare results of meteorological elements, such as 2 m temperature and precipitation generated by the HRADC with the gridded observation data from CMA, and surface temperature and specific humidity with Global LandData Assimilation System (GLDAS) output data from the National Aeronautics and Space Administration (NASA). We found that the satellite-derived GVF from MODIS increased over southeast China compared with the default model over the whole year. The simulated results of soil temperature, net radiation and surface energy flux from the HRADC are improved compared with the control simulation and are close to GLDAS outputs. The values of net radiation from HRADC are higher than the GLDAS outputs, and the differences in the simulations are large in the east region but are smaller in northwest China and on the Qinghai-Tibet Plateau. The spatial distribution of the sensible heat flux and the ground

  5. Spatially explicit modeling of particulate nutrient flux in Large global rivers

    Science.gov (United States)

    Cohen, S.; Kettner, A.; Mayorga, E.; Harrison, J. A.

    2017-12-01

    Water, sediment, nutrient and carbon fluxes along river networks have undergone considerable alterations in response to anthropogenic and climatic changes, with significant consequences to infrastructure, agriculture, water security, ecology and geomorphology worldwide. However, in a global setting, these changes in fluvial fluxes and their spatial and temporal characteristics are poorly constrained, due to the limited availability of continuous and long-term observations. We present results from a new global-scale particulate modeling framework (WBMsedNEWS) that combines the Global NEWS watershed nutrient export model with the spatially distributed WBMsed water and sediment model. We compare the model predictions against multiple observational datasets. The results indicate that the model is able to accurately predict particulate nutrient (Nitrogen, Phosphorus and Organic Carbon) fluxes on an annual time scale. Analysis of intra-basin nutrient dynamics and fluxes to global oceans is presented.

  6. iMS2Flux – a high–throughput processing tool for stable isotope labeled mass spectrometric data used for metabolic flux analysis

    Directory of Open Access Journals (Sweden)

    Poskar C Hart

    2012-11-01

    Full Text Available Abstract Background Metabolic flux analysis has become an established method in systems biology and functional genomics. The most common approach for determining intracellular metabolic fluxes is to utilize mass spectrometry in combination with stable isotope labeling experiments. However, before the mass spectrometric data can be used it has to be corrected for biases caused by naturally occurring stable isotopes, by the analytical technique(s employed, or by the biological sample itself. Finally the MS data and the labeling information it contains have to be assembled into a data format usable by flux analysis software (of which several dedicated packages exist. Currently the processing of mass spectrometric data is time-consuming and error-prone requiring peak by peak cut-and-paste analysis and manual curation. In order to facilitate high-throughput metabolic flux analysis, the automation of multiple steps in the analytical workflow is necessary. Results Here we describe iMS2Flux, software developed to automate, standardize and connect the data flow between mass spectrometric measurements and flux analysis programs. This tool streamlines the transfer of data from extraction via correction tools to 13C-Flux software by processing MS data from stable isotope labeling experiments. It allows the correction of large and heterogeneous MS datasets for the presence of naturally occurring stable isotopes, initial biomass and several mass spectrometry effects. Before and after data correction, several checks can be performed to ensure accurate data. The corrected data may be returned in a variety of formats including those used by metabolic flux analysis software such as 13CFLUX, OpenFLUX and 13CFLUX2. Conclusion iMS2Flux is a versatile, easy to use tool for the automated processing of mass spectrometric data containing isotope labeling information. It represents the core framework for a standardized workflow and data processing. Due to its flexibility

  7. Use of Principal Components Analysis to Explain Controls on Nutrient Fluxes to the Chesapeake Bay

    Science.gov (United States)

    Rice, K. C.; Mills, A. L.

    2017-12-01

    The Chesapeake Bay watershed, on the east coast of the United States, encompasses about 166,000-square kilometers (km2) of diverse land use, which includes a mixture of forested, agricultural, and developed land. The watershed is now managed under a Total Daily Maximum Load (TMDL), which requires implementation of management actions by 2025 that are sufficient to reduce nitrogen, phosphorus, and suspended-sediment fluxes to the Chesapeake Bay and restore the bay's water quality. We analyzed nutrient and sediment data along with land-use and climatic variables in nine sub watersheds to better understand the drivers of flux within the watershed and to provide relevant management implications. The nine sub watersheds range in area from 300 to 30,000 km2, and the analysis period was 1985-2014. The 31 variables specific to each sub watershed were highly statistically significantly correlated, so Principal Components Analysis was used to reduce the dimensionality of the dataset. The analysis revealed that about 80% of the variability in the whole dataset can be explained by discharge, flux, and concentration of nutrients and sediment. The first two principal components (PCs) explained about 68% of the total variance. PC1 loaded strongly on discharge and flux, and PC2 loaded on concentration. The PC scores of both PC1 and PC2 varied by season. Subsequent analysis of PC1 scores versus PC2 scores, broken out by sub watershed, revealed management implications. Some of the largest sub watersheds are largely driven by discharge, and consequently large fluxes. In contrast, some of the smaller sub watersheds are more variable in nutrient concentrations than discharge and flux. Our results suggest that, given no change in discharge, a reduction in nutrient flux to the streams in the smaller watersheds could result in a proportionately larger decrease in fluxes of nutrients down the river to the bay, than in the larger watersheds.

  8. A multimodal MRI dataset of professional chess players.

    Science.gov (United States)

    Li, Kaiming; Jiang, Jing; Qiu, Lihua; Yang, Xun; Huang, Xiaoqi; Lui, Su; Gong, Qiyong

    2015-01-01

    Chess is a good model to study high-level human brain functions such as spatial cognition, memory, planning, learning and problem solving. Recent studies have demonstrated that non-invasive MRI techniques are valuable for researchers to investigate the underlying neural mechanism of playing chess. For professional chess players (e.g., chess grand masters and masters or GM/Ms), what are the structural and functional alterations due to long-term professional practice, and how these alterations relate to behavior, are largely veiled. Here, we report a multimodal MRI dataset from 29 professional Chinese chess players (most of whom are GM/Ms), and 29 age matched novices. We hope that this dataset will provide researchers with new materials to further explore high-level human brain functions.

  9. Knowledge discovery with classification rules in a cardiovascular dataset.

    Science.gov (United States)

    Podgorelec, Vili; Kokol, Peter; Stiglic, Milojka Molan; Hericko, Marjan; Rozman, Ivan

    2005-12-01

    In this paper we study an evolutionary machine learning approach to data mining and knowledge discovery based on the induction of classification rules. A method for automatic rules induction called AREX using evolutionary induction of decision trees and automatic programming is introduced. The proposed algorithm is applied to a cardiovascular dataset consisting of different groups of attributes which should possibly reveal the presence of some specific cardiovascular problems in young patients. A case study is presented that shows the use of AREX for the classification of patients and for discovering possible new medical knowledge from the dataset. The defined knowledge discovery loop comprises a medical expert's assessment of induced rules to drive the evolution of rule sets towards more appropriate solutions. The final result is the discovery of a possible new medical knowledge in the field of pediatric cardiology.

  10. Augmented Reality Prototype for Visualizing Large Sensors’ Datasets

    Directory of Open Access Journals (Sweden)

    Folorunso Olufemi A.

    2011-04-01

    Full Text Available This paper addressed the development of an augmented reality (AR based scientific visualization system prototype that supports identification, localisation, and 3D visualisation of oil leakages sensors datasets. Sensors generates significant amount of multivariate datasets during normal and leak situations which made data exploration and visualisation daunting tasks. Therefore a model to manage such data and enhance computational support needed for effective explorations are developed in this paper. A challenge of this approach is to reduce the data inefficiency. This paper presented a model for computing information gain for each data attributes and determine a lead attribute.The computed lead attribute is then used for the development of an AR-based scientific visualization interface which automatically identifies, localises and visualizes all necessary data relevant to a particularly selected region of interest (ROI on the network. Necessary architectural system supports and the interface requirements for such visualizations are also presented.

  11. An integrated dataset for in silico drug discovery

    Directory of Open Access Journals (Sweden)

    Cockell Simon J

    2010-12-01

    Full Text Available Drug development is expensive and prone to failure. It is potentially much less risky and expensive to reuse a drug developed for one condition for treating a second disease, than it is to develop an entirely new compound. Systematic approaches to drug repositioning are needed to increase throughput and find candidates more reliably. Here we address this need with an integrated systems biology dataset, developed using the Ondex data integration platform, for the in silico discovery of new drug repositioning candidates. We demonstrate that the information in this dataset allows known repositioning examples to be discovered. We also propose a means of automating the search for new treatment indications of existing compounds.

  12. Flux flow and flux dynamics in high-Tc superconductors

    International Nuclear Information System (INIS)

    Bennett, L.H.; Turchinskaya, M.; Swartzendruber, L.J.; Roitburd, A.; Lundy, D.; Ritter, J.; Kaiser, D.L.

    1991-01-01

    Because high temperature superconductors, including BYCO and BSSCO, are type 2 superconductors with relatively low H(sub c 1) values and high H(sub c 2) values, they will be in a critical state for many of their applications. In the critical state, with the applied field between H(sub c 1) and H(sub c 2), flux lines have penetrated the material and can form a flux lattice and can be pinned by structural defects, chemical inhomogeneities, and impurities. A detailed knowledge of how flux penetrates the material and its behavior under the influence of applied fields and current flow, and the effect of material processing on these properties, is required in order to apply, and to improve the properties of these superconductors. When the applied field is changed rapidly, the time dependence of flux change can be divided into three regions, an initial region which occurs very rapidly, a second region in which the magnetization has a 1n(t) behavior, and a saturation region at very long times. A critical field is defined for depinning, H(sub c,p) as that field at which the hysteresis loop changes from irreversible to reversible. As a function of temperature, it is found that H(sub c,p) is well described by a power law with an exponent between 1.5 and 2.5. The behavior of H(sub c,p) for various materials and its relationship to flux flow and flux dynamics are discussed

  13. Application of Density Estimation Methods to Datasets from a Glider

    Science.gov (United States)

    2014-09-30

    humpback and sperm whales as well as different dolphin species. OBJECTIVES The objective of this research is to extend existing methods for cetacean...collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources...estimation from single sensor datasets. Required steps for a cue counting approach, where a cue has been defined as a clicking event (Küsel et al., 2011), to

  14. A review of continent scale hydrological datasets available for Africa

    OpenAIRE

    Bonsor, H.C.

    2010-01-01

    As rainfall becomes less reliable with predicted climate change the ability to assess the spatial and seasonal variations in groundwater availability on a large-scale (catchment and continent) is becoming increasingly important (Bates, et al. 2007; MacDonald et al. 2009). The scarcity of observed hydrological data, or difficulty in obtaining such data, within Africa means remotely sensed (RS) datasets must often be used to drive large-scale hydrological models. The different ap...

  15. Dataset of mitochondrial genome variants in oncocytic tumors

    Directory of Open Access Journals (Sweden)

    Lihua Lyu

    2018-04-01

    Full Text Available This dataset presents the mitochondrial genome variants associated with oncocytic tumors. These data were obtained by Sanger sequencing of the whole mitochondrial genomes of oncocytic tumors and the adjacent normal tissues from 32 patients. The mtDNA variants are identified after compared with the revised Cambridge sequence, excluding those defining haplogroups of our patients. The pathogenic prediction for the novel missense variants found in this study was performed with the Mitimpact 2 program.

  16. Soil chemistry in lithologically diverse datasets: the quartz dilution effect

    Science.gov (United States)

    Bern, Carleton R.

    2009-01-01

    National- and continental-scale soil geochemical datasets are likely to move our understanding of broad soil geochemistry patterns forward significantly. Patterns of chemistry and mineralogy delineated from these datasets are strongly influenced by the composition of the soil parent material, which itself is largely a function of lithology and particle size sorting. Such controls present a challenge by obscuring subtler patterns arising from subsequent pedogenic processes. Here the effect of quartz concentration is examined in moist-climate soils from a pilot dataset of the North American Soil Geochemical Landscapes Project. Due to variable and high quartz contents (6.2–81.7 wt.%), and its residual and inert nature in soil, quartz is demonstrated to influence broad patterns in soil chemistry. A dilution effect is observed whereby concentrations of various elements are significantly and strongly negatively correlated with quartz. Quartz content drives artificial positive correlations between concentrations of some elements and obscures negative correlations between others. Unadjusted soil data show the highly mobile base cations Ca, Mg, and Na to be often strongly positively correlated with intermediately mobile Al or Fe, and generally uncorrelated with the relatively immobile high-field-strength elements (HFS) Ti and Nb. Both patterns are contrary to broad expectations for soils being weathered and leached. After transforming bulk soil chemistry to a quartz-free basis, the base cations are generally uncorrelated with Al and Fe, and negative correlations generally emerge with the HFS elements. Quartz-free element data may be a useful tool for elucidating patterns of weathering or parent-material chemistry in large soil datasets.

  17. Dataset on records of Hericium erinaceus in Slovakia

    OpenAIRE

    Vladimír Kunca; Marek Čiliak

    2017-01-01

    The data presented in this article are related to the research article entitled ?Habitat preferences of Hericium erinaceus in Slovakia? (Kunca and ?iliak, 2016) [FUNECO607] [2]. The dataset include all available and unpublished data from Slovakia, besides the records from the same tree or stem. We compiled a database of records of collections by processing data from herbaria, personal records and communication with mycological activists. Data on altitude, tree species, host tree vital status,...

  18. Diffeomorphic Iterative Centroid Methods for Template Estimation on Large Datasets

    OpenAIRE

    Cury , Claire; Glaunès , Joan Alexis; Colliot , Olivier

    2014-01-01

    International audience; A common approach for analysis of anatomical variability relies on the stimation of a template representative of the population. The Large Deformation Diffeomorphic Metric Mapping is an attractive framework for that purpose. However, template estimation using LDDMM is computationally expensive, which is a limitation for the study of large datasets. This paper presents an iterative method which quickly provides a centroid of the population in the shape space. This centr...

  19. A Dataset from TIMSS to Examine the Relationship between Computer Use and Mathematics Achievement

    Science.gov (United States)

    Kadijevich, Djordje M.

    2015-01-01

    Because the relationship between computer use and achievement is still puzzling, there is a need to prepare and analyze good quality datasets on computer use and achievement. Such a dataset can be derived from TIMSS data. This paper describes how this dataset can be prepared. It also gives an example of how the dataset may be analyzed. The…

  20. An Analysis on Better Testing than Training Performances on the Iris Dataset

    NARCIS (Netherlands)

    Schutten, Marten; Wiering, Marco

    2016-01-01

    The Iris dataset is a well known dataset containing information on three different types of Iris flowers. A typical and popular method for solving classification problems on datasets such as the Iris set is the support vector machine (SVM). In order to do so the dataset is separated in a set used

  1. Parton Distributions based on a Maximally Consistent Dataset

    Science.gov (United States)

    Rojo, Juan

    2016-04-01

    The choice of data that enters a global QCD analysis can have a substantial impact on the resulting parton distributions and their predictions for collider observables. One of the main reasons for this has to do with the possible presence of inconsistencies, either internal within an experiment or external between different experiments. In order to assess the robustness of the global fit, different definitions of a conservative PDF set, that is, a PDF set based on a maximally consistent dataset, have been introduced. However, these approaches are typically affected by theory biases in the selection of the dataset. In this contribution, after a brief overview of recent NNPDF developments, we propose a new, fully objective, definition of a conservative PDF set, based on the Bayesian reweighting approach. Using the new NNPDF3.0 framework, we produce various conservative sets, which turn out to be mutually in agreement within the respective PDF uncertainties, as well as with the global fit. We explore some of their implications for LHC phenomenology, finding also good consistency with the global fit result. These results provide a non-trivial validation test of the new NNPDF3.0 fitting methodology, and indicate that possible inconsistencies in the fitted dataset do not affect substantially the global fit PDFs.

  2. New public dataset for spotting patterns in medieval document images

    Science.gov (United States)

    En, Sovann; Nicolas, Stéphane; Petitjean, Caroline; Jurie, Frédéric; Heutte, Laurent

    2017-01-01

    With advances in technology, a large part of our cultural heritage is becoming digitally available. In particular, in the field of historical document image analysis, there is now a growing need for indexing and data mining tools, thus allowing us to spot and retrieve the occurrences of an object of interest, called a pattern, in a large database of document images. Patterns may present some variability in terms of color, shape, or context, making the spotting of patterns a challenging task. Pattern spotting is a relatively new field of research, still hampered by the lack of available annotated resources. We present a new publicly available dataset named DocExplore dedicated to spotting patterns in historical document images. The dataset contains 1500 images and 1464 queries, and allows the evaluation of two tasks: image retrieval and pattern localization. A standardized benchmark protocol along with ad hoc metrics is provided for a fair comparison of the submitted approaches. We also provide some first results obtained with our baseline system on this new dataset, which show that there is room for improvement and that should encourage researchers of the document image analysis community to design new systems and submit improved results.

  3. Kernel-based discriminant feature extraction using a representative dataset

    Science.gov (United States)

    Li, Honglin; Sancho Gomez, Jose-Luis; Ahalt, Stanley C.

    2002-07-01

    Discriminant Feature Extraction (DFE) is widely recognized as an important pre-processing step in classification applications. Most DFE algorithms are linear and thus can only explore the linear discriminant information among the different classes. Recently, there has been several promising attempts to develop nonlinear DFE algorithms, among which is Kernel-based Feature Extraction (KFE). The efficacy of KFE has been experimentally verified by both synthetic data and real problems. However, KFE has some known limitations. First, KFE does not work well for strongly overlapped data. Second, KFE employs all of the training set samples during the feature extraction phase, which can result in significant computation when applied to very large datasets. Finally, KFE can result in overfitting. In this paper, we propose a substantial improvement to KFE that overcomes the above limitations by using a representative dataset, which consists of critical points that are generated from data-editing techniques and centroid points that are determined by using the Frequency Sensitive Competitive Learning (FSCL) algorithm. Experiments show that this new KFE algorithm performs well on significantly overlapped datasets, and it also reduces computational complexity. Further, by controlling the number of centroids, the overfitting problem can be effectively alleviated.

  4. Decoys Selection in Benchmarking Datasets: Overview and Perspectives

    Science.gov (United States)

    Réau, Manon; Langenfeld, Florent; Zagury, Jean-François; Lagarde, Nathalie; Montes, Matthieu

    2018-01-01

    Virtual Screening (VS) is designed to prospectively help identifying potential hits, i.e., compounds capable of interacting with a given target and potentially modulate its activity, out of large compound collections. Among the variety of methodologies, it is crucial to select the protocol that is the most adapted to the query/target system under study and that yields the most reliable output. To this aim, the performance of VS methods is commonly evaluated and compared by computing their ability to retrieve active compounds in benchmarking datasets. The benchmarking datasets contain a subset of known active compounds together with a subset of decoys, i.e., assumed non-active molecules. The composition of both the active and the decoy compounds subsets is critical to limit the biases in the evaluation of the VS methods. In this review, we focus on the selection of decoy compounds that has considerably changed over the years, from randomly selected compounds to highly customized or experimentally validated negative compounds. We first outline the evolution of decoys selection in benchmarking databases as well as current benchmarking databases that tend to minimize the introduction of biases, and secondly, we propose recommendations for the selection and the design of benchmarking datasets. PMID:29416509

  5. ENHANCED DATA DISCOVERABILITY FOR IN SITU HYPERSPECTRAL DATASETS

    Directory of Open Access Journals (Sweden)

    B. Rasaiah

    2016-06-01

    Full Text Available Field spectroscopic metadata is a central component in the quality assurance, reliability, and discoverability of hyperspectral data and the products derived from it. Cataloguing, mining, and interoperability of these datasets rely upon the robustness of metadata protocols for field spectroscopy, and on the software architecture to support the exchange of these datasets. Currently no standard for in situ spectroscopy data or metadata protocols exist. This inhibits the effective sharing of growing volumes of in situ spectroscopy datasets, to exploit the benefits of integrating with the evolving range of data sharing platforms. A core metadataset for field spectroscopy was introduced by Rasaiah et al., (2011-2015 with extended support for specific applications. This paper presents a prototype model for an OGC and ISO compliant platform-independent metadata discovery service aligned to the specific requirements of field spectroscopy. In this study, a proof-of-concept metadata catalogue has been described and deployed in a cloud-based architecture as a demonstration of an operationalized field spectroscopy metadata standard and web-based discovery service.

  6. Multiresolution persistent homology for excessively large biomolecular datasets

    Energy Technology Data Exchange (ETDEWEB)

    Xia, Kelin; Zhao, Zhixiong [Department of Mathematics, Michigan State University, East Lansing, Michigan 48824 (United States); Wei, Guo-Wei, E-mail: wei@math.msu.edu [Department of Mathematics, Michigan State University, East Lansing, Michigan 48824 (United States); Department of Electrical and Computer Engineering, Michigan State University, East Lansing, Michigan 48824 (United States); Department of Biochemistry and Molecular Biology, Michigan State University, East Lansing, Michigan 48824 (United States)

    2015-10-07

    Although persistent homology has emerged as a promising tool for the topological simplification of complex data, it is computationally intractable for large datasets. We introduce multiresolution persistent homology to handle excessively large datasets. We match the resolution with the scale of interest so as to represent large scale datasets with appropriate resolution. We utilize flexibility-rigidity index to access the topological connectivity of the data set and define a rigidity density for the filtration analysis. By appropriately tuning the resolution of the rigidity density, we are able to focus the topological lens on the scale of interest. The proposed multiresolution topological analysis is validated by a hexagonal fractal image which has three distinct scales. We further demonstrate the proposed method for extracting topological fingerprints from DNA molecules. In particular, the topological persistence of a virus capsid with 273 780 atoms is successfully analyzed which would otherwise be inaccessible to the normal point cloud method and unreliable by using coarse-grained multiscale persistent homology. The proposed method has also been successfully applied to the protein domain classification, which is the first time that persistent homology is used for practical protein domain analysis, to our knowledge. The proposed multiresolution topological method has potential applications in arbitrary data sets, such as social networks, biological networks, and graphs.

  7. Tissue-Based MRI Intensity Standardization: Application to Multicentric Datasets

    Directory of Open Access Journals (Sweden)

    Nicolas Robitaille

    2012-01-01

    Full Text Available Intensity standardization in MRI aims at correcting scanner-dependent intensity variations. Existing simple and robust techniques aim at matching the input image histogram onto a standard, while we think that standardization should aim at matching spatially corresponding tissue intensities. In this study, we present a novel automatic technique, called STI for STandardization of Intensities, which not only shares the simplicity and robustness of histogram-matching techniques, but also incorporates tissue spatial intensity information. STI uses joint intensity histograms to determine intensity correspondence in each tissue between the input and standard images. We compared STI to an existing histogram-matching technique on two multicentric datasets, Pilot E-ADNI and ADNI, by measuring the intensity error with respect to the standard image after performing nonlinear registration. The Pilot E-ADNI dataset consisted in 3 subjects each scanned in 7 different sites. The ADNI dataset consisted in 795 subjects scanned in more than 50 different sites. STI was superior to the histogram-matching technique, showing significantly better intensity matching for the brain white matter with respect to the standard image.

  8. Exploring massive, genome scale datasets with the genometricorr package

    KAUST Repository

    Favorov, Alexander; Mularoni, Loris; Cope, Leslie M.; Medvedeva, Yulia; Mironov, Andrey A.; Makeev, Vsevolod J.; Wheelan, Sarah J.

    2012-01-01

    We have created a statistically grounded tool for determining the correlation of genomewide data with other datasets or known biological features, intended to guide biological exploration of high-dimensional datasets, rather than providing immediate answers. The software enables several biologically motivated approaches to these data and here we describe the rationale and implementation for each approach. Our models and statistics are implemented in an R package that efficiently calculates the spatial correlation between two sets of genomic intervals (data and/or annotated features), for use as a metric of functional interaction. The software handles any type of pointwise or interval data and instead of running analyses with predefined metrics, it computes the significance and direction of several types of spatial association; this is intended to suggest potentially relevant relationships between the datasets. Availability and implementation: The package, GenometriCorr, can be freely downloaded at http://genometricorr.sourceforge.net/. Installation guidelines and examples are available from the sourceforge repository. The package is pending submission to Bioconductor. © 2012 Favorov et al.

  9. Image segmentation evaluation for very-large datasets

    Science.gov (United States)

    Reeves, Anthony P.; Liu, Shuang; Xie, Yiting

    2016-03-01

    With the advent of modern machine learning methods and fully automated image analysis there is a need for very large image datasets having documented segmentations for both computer algorithm training and evaluation. Current approaches of visual inspection and manual markings do not scale well to big data. We present a new approach that depends on fully automated algorithm outcomes for segmentation documentation, requires no manual marking, and provides quantitative evaluation for computer algorithms. The documentation of new image segmentations and new algorithm outcomes are achieved by visual inspection. The burden of visual inspection on large datasets is minimized by (a) customized visualizations for rapid review and (b) reducing the number of cases to be reviewed through analysis of quantitative segmentation evaluation. This method has been applied to a dataset of 7,440 whole-lung CT images for 6 different segmentation algorithms designed to fully automatically facilitate the measurement of a number of very important quantitative image biomarkers. The results indicate that we could achieve 93% to 99% successful segmentation for these algorithms on this relatively large image database. The presented evaluation method may be scaled to much larger image databases.

  10. Exploring massive, genome scale datasets with the genometricorr package

    KAUST Repository

    Favorov, Alexander

    2012-05-31

    We have created a statistically grounded tool for determining the correlation of genomewide data with other datasets or known biological features, intended to guide biological exploration of high-dimensional datasets, rather than providing immediate answers. The software enables several biologically motivated approaches to these data and here we describe the rationale and implementation for each approach. Our models and statistics are implemented in an R package that efficiently calculates the spatial correlation between two sets of genomic intervals (data and/or annotated features), for use as a metric of functional interaction. The software handles any type of pointwise or interval data and instead of running analyses with predefined metrics, it computes the significance and direction of several types of spatial association; this is intended to suggest potentially relevant relationships between the datasets. Availability and implementation: The package, GenometriCorr, can be freely downloaded at http://genometricorr.sourceforge.net/. Installation guidelines and examples are available from the sourceforge repository. The package is pending submission to Bioconductor. © 2012 Favorov et al.

  11. Principal Component Analysis of Process Datasets with Missing Values

    Directory of Open Access Journals (Sweden)

    Kristen A. Severson

    2017-07-01

    Full Text Available Datasets with missing values arising from causes such as sensor failure, inconsistent sampling rates, and merging data from different systems are common in the process industry. Methods for handling missing data typically operate during data pre-processing, but can also occur during model building. This article considers missing data within the context of principal component analysis (PCA, which is a method originally developed for complete data that has widespread industrial application in multivariate statistical process control. Due to the prevalence of missing data and the success of PCA for handling complete data, several PCA algorithms that can act on incomplete data have been proposed. Here, algorithms for applying PCA to datasets with missing values are reviewed. A case study is presented to demonstrate the performance of the algorithms and suggestions are made with respect to choosing which algorithm is most appropriate for particular settings. An alternating algorithm based on the singular value decomposition achieved the best results in the majority of test cases involving process datasets.

  12. A cross-country Exchange Market Pressure (EMP dataset

    Directory of Open Access Journals (Sweden)

    Mohit Desai

    2017-06-01

    Full Text Available The data presented in this article are related to the research article titled - “An exchange market pressure measure for cross country analysis” (Patnaik et al. [1]. In this article, we present the dataset for Exchange Market Pressure values (EMP for 139 countries along with their conversion factors, ρ (rho. Exchange Market Pressure, expressed in percentage change in exchange rate, measures the change in exchange rate that would have taken place had the central bank not intervened. The conversion factor ρ can interpreted as the change in exchange rate associated with $1 billion of intervention. Estimates of conversion factor ρ allow us to calculate a monthly time series of EMP for 139 countries. Additionally, the dataset contains the 68% confidence interval (high and low values for the point estimates of ρ’s. Using the standard errors of estimates of ρ’s, we obtain one sigma intervals around mean estimates of EMP values. These values are also reported in the dataset.

  13. Validation of a Meteosat Second Generation solar radiation dataset over the northeastern Iberian Peninsula

    Directory of Open Access Journals (Sweden)

    J. Cristóbal

    2013-01-01

    Full Text Available Solar radiation plays a key role in the Earth's energy balance and is used as an essential input data in radiation-based evapotranspiration (ET models. Accurate gridded solar radiation data at high spatial and temporal resolution are needed to retrieve ET over large domains. In this work we present an evaluation at hourly, daily and monthly time steps and regional scale (Catalonia, NE Iberian Peninsula of a satellite-based solar radiation product developed by the Land Surface Analysis Satellite Application Facility (LSA SAF using data from the Meteosat Second Generation (MSG Spinning Enhanced Visible and Infrared Imager (SEVIRI. Product performance and accuracy were evaluated for datasets segmented into two terrain classes (flat and hilly areas and two atmospheric conditions (clear and cloudy sky, as well as for the full dataset as a whole. Evaluation against measurements made with ground-based pyranometers yielded good results in flat areas with an averaged model RMSE of 65 W m−2 (19%, 34 W m−2 (9.7% and 21 W m−2 (5.6%, for hourly, daily and monthly-averaged solar radiation and including clear and cloudy sky conditions and snow or ice cover. Hilly areas yielded intermediate results with an averaged model RMSE (root mean square error of 89 W m−2 (27%, 48 W m−2 (14.5% and 32 W m−2 (9.3%, for hourly, daily and monthly time steps, suggesting the need of further improvements (e.g., terrain corrections required for retrieving localized variability in solar radiation in these areas. According to the literature, the LSA SAF solar radiation product appears to have sufficient accuracy to serve as a useful and operative input to evaporative flux retrieval models.

  14. Thermality of the Hawking flux

    Energy Technology Data Exchange (ETDEWEB)

    Visser, Matt [School of Mathematics, Statistics, and Operations Research,Victoria University of Wellington, PO Box 600, Wellington 6140 (New Zealand)

    2015-07-03

    Is the Hawking flux “thermal”? Unfortunately, the answer to this seemingly innocent question depends on a number of often unstated, but quite crucial, technical assumptions built into modern (mis-)interpretations of the word “thermal”. The original 1850’s notions of thermality — based on classical thermodynamic reasoning applied to idealized “black bodies” or “lamp black surfaces” — when supplemented by specific basic quantum ideas from the early 1900’s, immediately led to the notion of the black-body spectrum, (the Planck-shaped spectrum), but without any specific assumptions or conclusions regarding correlations between the quanta. Many (not all) modern authors (often implicitly and unintentionally) add an extra, quite unnecessary, assumption that there are no correlations in the black-body radiation; but such usage is profoundly ahistorical and dangerously misleading. Specifically, the Hawking flux from an evaporating black hole, (just like the radiation flux from a leaky furnace or a burning lump of coal), is only approximately Planck-shaped over an explicitly bounded range of frequencies. Standard physics (phase space and adiabaticity effects) explicitly bound the frequency range over which the Hawking flux is approximately Planck-shaped from both above and below — the Hawking flux is certainly not exactly Planckian, and there is no compelling physics reason to assume the Hawking photons are uncorrelated.

  15. Thermality of the Hawking flux

    International Nuclear Information System (INIS)

    Visser, Matt

    2015-01-01

    Is the Hawking flux “thermal”? Unfortunately, the answer to this seemingly innocent question depends on a number of often unstated, but quite crucial, technical assumptions built into modern (mis-)interpretations of the word “thermal”. The original 1850’s notions of thermality — based on classical thermodynamic reasoning applied to idealized “black bodies” or “lamp black surfaces” — when supplemented by specific basic quantum ideas from the early 1900’s, immediately led to the notion of the black-body spectrum, (the Planck-shaped spectrum), but without any specific assumptions or conclusions regarding correlations between the quanta. Many (not all) modern authors (often implicitly and unintentionally) add an extra, quite unnecessary, assumption that there are no correlations in the black-body radiation; but such usage is profoundly ahistorical and dangerously misleading. Specifically, the Hawking flux from an evaporating black hole, (just like the radiation flux from a leaky furnace or a burning lump of coal), is only approximately Planck-shaped over an explicitly bounded range of frequencies. Standard physics (phase space and adiabaticity effects) explicitly bound the frequency range over which the Hawking flux is approximately Planck-shaped from both above and below — the Hawking flux is certainly not exactly Planckian, and there is no compelling physics reason to assume the Hawking photons are uncorrelated.

  16. Physics of magnetic flux tubes

    CERN Document Server

    Ryutova, Margarita

    2015-01-01

    This book is the first account of the physics of magnetic flux tubes from their fundamental properties to collective phenomena in an ensembles of flux tubes. The physics of magnetic flux tubes is absolutely vital for understanding fundamental physical processes in the solar atmosphere shaped and governed by magnetic fields. High-resolution and high cadence observations from recent space and  ground-based instruments taken simultaneously at different heights and temperatures not only show the ubiquity of filamentary structure formation but also allow to study how various events are interconnected by system of magnetic flux tubes. The book covers both theory and observations. Theoretical models presented in analytical and phenomenological forms are tailored for practical applications. These are welded with state-of-the-art observations from early decisive ones to the most recent data that open a new phase-space for exploring the Sun and sun-like stars. Concept of magnetic flux tubes is central to various magn...

  17. The Role of Datasets on Scientific Influence within Conflict Research

    Science.gov (United States)

    Van Holt, Tracy; Johnson, Jeffery C.; Moates, Shiloh; Carley, Kathleen M.

    2016-01-01

    We inductively tested if a coherent field of inquiry in human conflict research emerged in an analysis of published research involving “conflict” in the Web of Science (WoS) over a 66-year period (1945–2011). We created a citation network that linked the 62,504 WoS records and their cited literature. We performed a critical path analysis (CPA), a specialized social network analysis on this citation network (~1.5 million works), to highlight the main contributions in conflict research and to test if research on conflict has in fact evolved to represent a coherent field of inquiry. Out of this vast dataset, 49 academic works were highlighted by the CPA suggesting a coherent field of inquiry; which means that researchers in the field acknowledge seminal contributions and share a common knowledge base. Other conflict concepts that were also analyzed—such as interpersonal conflict or conflict among pharmaceuticals, for example, did not form their own CP. A single path formed, meaning that there was a cohesive set of ideas that built upon previous research. This is in contrast to a main path analysis of conflict from 1957–1971 where ideas didn’t persist in that multiple paths existed and died or emerged reflecting lack of scientific coherence (Carley, Hummon, and Harty, 1993). The critical path consisted of a number of key features: 1) Concepts that built throughout include the notion that resource availability drives conflict, which emerged in the 1960s-1990s and continued on until 2011. More recent intrastate studies that focused on inequalities emerged from interstate studies on the democracy of peace earlier on the path. 2) Recent research on the path focused on forecasting conflict, which depends on well-developed metrics and theories to model. 3) We used keyword analysis to independently show how the CP was topically linked (i.e., through democracy, modeling, resources, and geography). Publically available conflict datasets developed early on helped

  18. The Role of Datasets on Scientific Influence within Conflict Research.

    Directory of Open Access Journals (Sweden)

    Tracy Van Holt

    Full Text Available We inductively tested if a coherent field of inquiry in human conflict research emerged in an analysis of published research involving "conflict" in the Web of Science (WoS over a 66-year period (1945-2011. We created a citation network that linked the 62,504 WoS records and their cited literature. We performed a critical path analysis (CPA, a specialized social network analysis on this citation network (~1.5 million works, to highlight the main contributions in conflict research and to test if research on conflict has in fact evolved to represent a coherent field of inquiry. Out of this vast dataset, 49 academic works were highlighted by the CPA suggesting a coherent field of inquiry; which means that researchers in the field acknowledge seminal contributions and share a common knowledge base. Other conflict concepts that were also analyzed-such as interpersonal conflict or conflict among pharmaceuticals, for example, did not form their own CP. A single path formed, meaning that there was a cohesive set of ideas that built upon previous research. This is in contrast to a main path analysis of conflict from 1957-1971 where ideas didn't persist in that multiple paths existed and died or emerged reflecting lack of scientific coherence (Carley, Hummon, and Harty, 1993. The critical path consisted of a number of key features: 1 Concepts that built throughout include the notion that resource availability drives conflict, which emerged in the 1960s-1990s and continued on until 2011. More recent intrastate studies that focused on inequalities emerged from interstate studies on the democracy of peace earlier on the path. 2 Recent research on the path focused on forecasting conflict, which depends on well-developed metrics and theories to model. 3 We used keyword analysis to independently show how the CP was topically linked (i.e., through democracy, modeling, resources, and geography. Publically available conflict datasets developed early on helped

  19. The Role of Datasets on Scientific Influence within Conflict Research.

    Science.gov (United States)

    Van Holt, Tracy; Johnson, Jeffery C; Moates, Shiloh; Carley, Kathleen M

    2016-01-01

    We inductively tested if a coherent field of inquiry in human conflict research emerged in an analysis of published research involving "conflict" in the Web of Science (WoS) over a 66-year period (1945-2011). We created a citation network that linked the 62,504 WoS records and their cited literature. We performed a critical path analysis (CPA), a specialized social network analysis on this citation network (~1.5 million works), to highlight the main contributions in conflict research and to test if research on conflict has in fact evolved to represent a coherent field of inquiry. Out of this vast dataset, 49 academic works were highlighted by the CPA suggesting a coherent field of inquiry; which means that researchers in the field acknowledge seminal contributions and share a common knowledge base. Other conflict concepts that were also analyzed-such as interpersonal conflict or conflict among pharmaceuticals, for example, did not form their own CP. A single path formed, meaning that there was a cohesive set of ideas that built upon previous research. This is in contrast to a main path analysis of conflict from 1957-1971 where ideas didn't persist in that multiple paths existed and died or emerged reflecting lack of scientific coherence (Carley, Hummon, and Harty, 1993). The critical path consisted of a number of key features: 1) Concepts that built throughout include the notion that resource availability drives conflict, which emerged in the 1960s-1990s and continued on until 2011. More recent intrastate studies that focused on inequalities emerged from interstate studies on the democracy of peace earlier on the path. 2) Recent research on the path focused on forecasting conflict, which depends on well-developed metrics and theories to model. 3) We used keyword analysis to independently show how the CP was topically linked (i.e., through democracy, modeling, resources, and geography). Publically available conflict datasets developed early on helped shape the

  20. Flux driven turbulence in tokamaks

    International Nuclear Information System (INIS)

    Garbet, X.; Ghendrih, P.; Ottaviani, M.; Sarazin, Y.; Beyer, P.; Benkadda, S.; Waltz, R.E.

    1999-01-01

    This work deals with tokamak plasma turbulence in the case where fluxes are fixed and profiles are allowed to fluctuate. These systems are intermittent. In particular, radially propagating fronts, are usually observed over a broad range of time and spatial scales. The existence of these fronts provide a way to understand the fast transport events sometimes observed in tokamaks. It is also shown that the confinement scaling law can still be of the gyroBohm type in spite of these large scale transport events. Some departure from the gyroBohm prediction is observed at low flux, i.e. when the gradients are close to the instability threshold. Finally, it is found that the diffusivity is not the same for a turbulence calculated at fixed flux than at fixed temperature gradient, with the same time averaged profile. (author)

  1. Methane flux from boreal peatlands

    International Nuclear Information System (INIS)

    Crill, P.; Bartlett, K.; Roulet, N.

    1992-01-01

    The peatlands in the boreal zone (roughly 45 deg - 60 degN) store a significant reservoir of carbon, much of which is potentially available for exchange with the atmosphere. The anaerobic conditions that cause these soils to accumulate carbon also makes wet, boreal peatlands significant sources of methane to the global troposphere. It is estimated that boreal wetlands contribute approximately 19.5 Tg methane per year. The data available on the magnitude of boreal methane emissions have rapidly accumulated in the past twenty years. This paper offers a short review of the flux measured (with range roughly 1 - 2000 mg methane/m2d), considers environmental controls of the flux and briefly discusses how climate change might affect future fluxes

  2. Wide range neutron flux monitor

    International Nuclear Information System (INIS)

    Endo, Yorimasa; Fukushima, Toshiki.

    1983-01-01

    Purpose: To provide a wide range neutron-flux monitor adapted such that the flux monitoring function and alarming function can automatically by shifted from pulse counting system to cambel method system. Constitution: A wide range neutron-flux monitor comprises (la) pulse counting system and (lb) cambel-method system for inputting detection signals from neutron detectors and separating them into signals for the pulse measuring system and the cambel measuring system, (2) overlap detection and calculation circuit for detecting the existence of the overlap of two output signals from the (la) and (lb) systems, and (3) trip circuit for judging the abnormal state of neutron detectors upon input of the detection signals. (Seki, T.)

  3. High heat flux facility GLADIS

    International Nuclear Information System (INIS)

    Greuner, H.; Boeswirth, B.; Boscary, J.; McNeely, P.

    2007-01-01

    The new ion beam facility GLADIS started the operation at IPP Garching. The facility is equipped with two individual 1.1 MW power ion sources for testing actively cooled plasma facing components under high heat fluxes. Each ion source generates heat loads between 3 and 55 MW/m 2 with a beam diameter of 70 mm at the target position. These parameters allow effective testing from probes to large components up to 2 m length. The high heat flux allows the target to be installed inclined to the beam and thus increases the heated surface length up to 200 mm for a heat flux of 15 MW/m 2 in the standard operating regime. Thus the facility has the potential capability for testing of full scale ITER divertor targets. Heat load tests on the WENDELSTEIN 7-X pre-series divertor targets have been successfully started. These tests will validate the design and manufacturing for the production of 950 elements

  4. Heat flux driven ion turbulence

    International Nuclear Information System (INIS)

    Garbet, X.

    1998-01-01

    This work is an analysis of an ion turbulence in a tokamak in the case where the thermal flux is fixed and the temperature profile is allowed to fluctuate. The system exhibits some features of Self-Organized Critical systems. In particular, avalanches are observed. Also the frequency spectrum of the thermal flux exhibits a structure similar to the one of a sand pile automaton, including a 1/f behavior. However, the time average temperature profile is found to be supercritical, i.e. the temperature gradient stays above the critical value. Moreover, the heat diffusivity is lower for a turbulence calculated at fixed flux than a fixed temperature gradient, with the same time average temperature. This behavior is attributed to a stabilizing effect of avalanches. (author)

  5. Ideal flux field dielectric concentrators.

    Science.gov (United States)

    García-Botella, Angel

    2011-10-01

    The concept of the vector flux field was first introduced as a photometrical theory and later developed in the field of nonimaging optics; it has provided new perspectives in the design of concentrators, overcoming standard ray tracing techniques. The flux field method has shown that reflective concentrators with the geometry of the field lines achieve the theoretical limit of concentration. In this paper we study the role of surfaces orthogonal to the field vector J. For rotationally symmetric systems J is orthogonal to its curl, and then a family of surfaces orthogonal to the lines of J exists, which can be called the family of surfaces of constant pseudopotential. Using the concept of the flux tube, it is possible to demonstrate that refractive concentrators with the shape of these pseudopotential surfaces achieve the theoretical limit of concentration.

  6. Flux flow and flux creep in thick films of YBCO. [Y-Ba-Cu-O

    Energy Technology Data Exchange (ETDEWEB)

    Rickets, J.; Vinen, W.F.; Abell, J.S.; Shields, T.C. (Superconductivity Research Group, Univ. of Birmingham (United Kingdom))

    1991-12-01

    The results are described of new experiments designed to study flux creep and flux flow along a single flux percolation path in thick films of YBCO. The flux flow regime is studied by a four-point resistive technique using pulsed currents, and the flux creep regime by observing the rate at which flux enters a superconducting loop in parallel with the resistance that is associated with the flux percolation path. (orig.).

  7. Assessment of radiation damage behaviour in a large collection of empirically optimized datasets highlights the importance of unmeasured complicating effects

    International Nuclear Information System (INIS)

    Krojer, Tobias; Delft, Frank von

    2011-01-01

    A retrospective analysis of radiation damage behaviour in a statistically significant number of real-life datasets is presented, in order to gauge the importance of the complications not yet measured or rigorously evaluated in current experiments, and the challenges that remain before radiation damage can be considered a problem solved in practice. The radiation damage behaviour in 43 datasets of 34 different proteins collected over a year was examined, in order to gauge the reliability of decay metrics in practical situations, and to assess how these datasets, optimized only empirically for decay, would have benefited from the precise and automatic prediction of decay now possible with the programs RADDOSE [Murray, Garman & Ravelli (2004 ▶). J. Appl. Cryst.37, 513–522] and BEST [Bourenkov & Popov (2010 ▶). Acta Cryst. D66, 409–419]. The results indicate that in routine practice the diffraction experiment is not yet characterized well enough to support such precise predictions, as these depend fundamentally on three interrelated variables which cannot yet be determined robustly and practically: the flux density distribution of the beam; the exact crystal volume; the sensitivity of the crystal to dose. The former two are not satisfactorily approximated from typical beamline information such as nominal beam size and transmission, or two-dimensional images of the beam and crystal; the discrepancies are particularly marked when using microfocus beams (<20 µm). Empirically monitoring decay with the dataset scaling B factor (Bourenkov & Popov, 2010 ▶) appears more robust but is complicated by anisotropic and/or low-resolution diffraction. These observations serve to delineate the challenges, scientific and logistic, that remain to be addressed if tools for managing radiation damage in practical data collection are to be conveniently robust enough to be useful in real time

  8. The flux database concerted action

    International Nuclear Information System (INIS)

    Mitchell, N.G.; Donnelly, C.E.

    1999-01-01

    This paper summarizes the background to the UIR action on the development of a flux database for radionuclide transfer in soil-plant systems. The action is discussed in terms of the objectives, the deliverables and the progress achieved so far by the flux database working group. The paper describes the background to the current initiative and outlines specific features of the database and supporting documentation. Particular emphasis is placed on the proforma used for data entry, on the database help file and on the approach adopted to indicate data quality. Refs. 3 (author)

  9. Strong low-pass filtering effects on water vapour flux measurements with closed-path eddy correlation systems

    DEFF Research Database (Denmark)

    Ibrom, Andreas; Dellwik, Ebba; Flyvbjerg, Henrik K.

    2007-01-01

    datasets for this substantial measurement error. In contrast to earlier studies, a large number of spectra and raw data have been used in the analysis to define the low-pass filtering characteristic of the EC system. This revealed that the cut-off frequency of the closed-path EC system for water vapour......Turbulent water vapour fluxes measured with closed-path eddy correlation (EC) systems are unintentionally low-pass filtered by the system in a manner that varies with environmental conditions. Why and how is described here. So is the practical method that systematically corrects long-term flux...... concentration measurements decreases exponentially with increasing relative humidity. After correction for this unintended filtering, the fluxes are consistent with CO2 and H2O fluxes that were measured with an open-path sensor at the same time. The correction of water vapour flux measurements over a Beech...

  10. Animated analysis of geoscientific datasets: An interactive graphical application

    Science.gov (United States)

    Morse, Peter; Reading, Anya; Lueg, Christopher

    2017-12-01

    Geoscientists are required to analyze and draw conclusions from increasingly large volumes of data. There is a need to recognise and characterise features and changing patterns of Earth observables within such large datasets. It is also necessary to identify significant subsets of the data for more detailed analysis. We present an innovative, interactive software tool and workflow to visualise, characterise, sample and tag large geoscientific datasets from both local and cloud-based repositories. It uses an animated interface and human-computer interaction to utilise the capacity of human expert observers to identify features via enhanced visual analytics. 'Tagger' enables users to analyze datasets that are too large in volume to be drawn legibly on a reasonable number of single static plots. Users interact with the moving graphical display, tagging data ranges of interest for subsequent attention. The tool provides a rapid pre-pass process using fast GPU-based OpenGL graphics and data-handling and is coded in the Quartz Composer visual programing language (VPL) on Mac OSX. It makes use of interoperable data formats, and cloud-based (or local) data storage and compute. In a case study, Tagger was used to characterise a decade (2000-2009) of data recorded by the Cape Sorell Waverider Buoy, located approximately 10 km off the west coast of Tasmania, Australia. These data serve as a proxy for the understanding of Southern Ocean storminess, which has both local and global implications. This example shows use of the tool to identify and characterise 4 different types of storm and non-storm events during this time. Events characterised in this way are compared with conventional analysis, noting advantages and limitations of data analysis using animation and human interaction. Tagger provides a new ability to make use of humans as feature detectors in computer-based analysis of large-volume geosciences and other data.

  11. Designing the colorectal cancer core dataset in Iran

    Directory of Open Access Journals (Sweden)

    Sara Dorri

    2017-01-01

    Full Text Available Background: There is no need to explain the importance of collection, recording and analyzing the information of disease in any health organization. In this regard, systematic design of standard data sets can be helpful to record uniform and consistent information. It can create interoperability between health care systems. The main purpose of this study was design the core dataset to record colorectal cancer information in Iran. Methods: For the design of the colorectal cancer core data set, a combination of literature review and expert consensus were used. In the first phase, the draft of the data set was designed based on colorectal cancer literature review and comparative studies. Then, in the second phase, this data set was evaluated by experts from different discipline such as medical informatics, oncology and surgery. Their comments and opinion were taken. In the third phase refined data set, was evaluated again by experts and eventually data set was proposed. Results: In first phase, based on the literature review, a draft set of 85 data elements was designed. In the second phase this data set was evaluated by experts and supplementary information was offered by professionals in subgroups especially in treatment part. In this phase the number of elements totally were arrived to 93 numbers. In the third phase, evaluation was conducted by experts and finally this dataset was designed in five main parts including: demographic information, diagnostic information, treatment information, clinical status assessment information, and clinical trial information. Conclusion: In this study the comprehensive core data set of colorectal cancer was designed. This dataset in the field of collecting colorectal cancer information can be useful through facilitating exchange of health information. Designing such data set for similar disease can help providers to collect standard data from patients and can accelerate retrieval from storage systems.

  12. FTSPlot: fast time series visualization for large datasets.

    Directory of Open Access Journals (Sweden)

    Michael Riss

    Full Text Available The analysis of electrophysiological recordings often involves visual inspection of time series data to locate specific experiment epochs, mask artifacts, and verify the results of signal processing steps, such as filtering or spike detection. Long-term experiments with continuous data acquisition generate large amounts of data. Rapid browsing through these massive datasets poses a challenge to conventional data plotting software because the plotting time increases proportionately to the increase in the volume of data. This paper presents FTSPlot, which is a visualization concept for large-scale time series datasets using techniques from the field of high performance computer graphics, such as hierarchic level of detail and out-of-core data handling. In a preprocessing step, time series data, event, and interval annotations are converted into an optimized data format, which then permits fast, interactive visualization. The preprocessing step has a computational complexity of O(n x log(N; the visualization itself can be done with a complexity of O(1 and is therefore independent of the amount of data. A demonstration prototype has been implemented and benchmarks show that the technology is capable of displaying large amounts of time series data, event, and interval annotations lag-free with < 20 ms ms. The current 64-bit implementation theoretically supports datasets with up to 2(64 bytes, on the x86_64 architecture currently up to 2(48 bytes are supported, and benchmarks have been conducted with 2(40 bytes/1 TiB or 1.3 x 10(11 double precision samples. The presented software is freely available and can be included as a Qt GUI component in future software projects, providing a standard visualization method for long-term electrophysiological experiments.

  13. A synthetic dataset for evaluating soft and hard fusion algorithms

    Science.gov (United States)

    Graham, Jacob L.; Hall, David L.; Rimland, Jeffrey

    2011-06-01

    There is an emerging demand for the development of data fusion techniques and algorithms that are capable of combining conventional "hard" sensor inputs such as video, radar, and multispectral sensor data with "soft" data including textual situation reports, open-source web information, and "hard/soft" data such as image or video data that includes human-generated annotations. New techniques that assist in sense-making over a wide range of vastly heterogeneous sources are critical to improving tactical situational awareness in counterinsurgency (COIN) and other asymmetric warfare situations. A major challenge in this area is the lack of realistic datasets available for test and evaluation of such algorithms. While "soft" message sets exist, they tend to be of limited use for data fusion applications due to the lack of critical message pedigree and other metadata. They also lack corresponding hard sensor data that presents reasonable "fusion opportunities" to evaluate the ability to make connections and inferences that span the soft and hard data sets. This paper outlines the design methodologies, content, and some potential use cases of a COIN-based synthetic soft and hard dataset created under a United States Multi-disciplinary University Research Initiative (MURI) program funded by the U.S. Army Research Office (ARO). The dataset includes realistic synthetic reports from a variety of sources, corresponding synthetic hard data, and an extensive supporting database that maintains "ground truth" through logical grouping of related data into "vignettes." The supporting database also maintains the pedigree of messages and other critical metadata.

  14. Simple models with ALICE fluxes

    CERN Document Server

    Striet, J

    2000-01-01

    We introduce two simple models which feature an Alice electrodynamics phase. In a well defined sense the Alice flux solutions we obtain in these models obey first order equations similar to those of the Nielsen-Olesen fluxtube in the abelian higgs model in the Bogomol'nyi limit. Some numerical solutions are presented as well.

  15. Identifying frauds and anomalies in Medicare-B dataset.

    Science.gov (United States)

    Jiwon Seo; Mendelevitch, Ofer

    2017-07-01

    Healthcare industry is growing at a rapid rate to reach a market value of $7 trillion dollars world wide. At the same time, fraud in healthcare is becoming a serious problem, amounting to 5% of the total healthcare spending, or $100 billion dollars each year in US. Manually detecting healthcare fraud requires much effort. Recently, machine learning and data mining techniques are applied to automatically detect healthcare frauds. This paper proposes a novel PageRank-based algorithm to detect healthcare frauds and anomalies. We apply the algorithm to Medicare-B dataset, a real-life data with 10 million healthcare insurance claims. The algorithm successfully identifies tens of previously unreported anomalies.

  16. Power analysis dataset for QCA based multiplexer circuits

    Directory of Open Access Journals (Sweden)

    Md. Abdullah-Al-Shafi

    2017-04-01

    Full Text Available Power consumption in irreversible QCA logic circuits is a vital and a major issue; however in the practical cases, this focus is mostly omitted.The complete power depletion dataset of different QCA multiplexers have been worked out in this paper. At −271.15 °C temperature, the depletion is evaluated under three separate tunneling energy levels. All the circuits are designed with QCADesigner, a broadly used simulation engine and QCAPro tool has been applied for estimating the power dissipation.

  17. Equalizing imbalanced imprecise datasets for genetic fuzzy classifiers

    Directory of Open Access Journals (Sweden)

    AnaM. Palacios

    2012-04-01

    Full Text Available Determining whether an imprecise dataset is imbalanced is not immediate. The vagueness in the data causes that the prior probabilities of the classes are not precisely known, and therefore the degree of imbalance can also be uncertain. In this paper we propose suitable extensions of different resampling algorithms that can be applied to interval valued, multi-labelled data. By means of these extended preprocessing algorithms, certain classification systems designed for minimizing the fraction of misclassifications are able to produce knowledge bases that are also adequate under common metrics for imbalanced classification.

  18. Scientific Datasets: Discovery and Aggregation for Semantic Interpretation.

    Science.gov (United States)

    Lopez, L. A.; Scott, S.; Khalsa, S. J. S.; Duerr, R.

    2015-12-01

    One of the biggest challenges that interdisciplinary researchers face is finding suitable datasets in order to advance their science; this problem remains consistent across multiple disciplines. A surprising number of scientists, when asked what tool they use for data discovery, reply "Google", which is an acceptable solution in some cases but not even Google can find -or cares to compile- all the data that's relevant for science and particularly geo sciences. If a dataset is not discoverable through a well known search provider it will remain dark data to the scientific world.For the past year, BCube, an EarthCube Building Block project, has been developing, testing and deploying a technology stack capable of data discovery at web-scale using the ultimate dataset: The Internet. This stack has 2 principal components, a web-scale crawling infrastructure and a semantic aggregator. The web-crawler is a modified version of Apache Nutch (the originator of Hadoop and other big data technologies) that has been improved and tailored for data and data service discovery. The second component is semantic aggregation, carried out by a python-based workflow that extracts valuable metadata and stores it in the form of triples through the use semantic technologies.While implementing the BCube stack we have run into several challenges such as a) scaling the project to cover big portions of the Internet at a reasonable cost, b) making sense of very diverse and non-homogeneous data, and lastly, c) extracting facts about these datasets using semantic technologies in order to make them usable for the geosciences community. Despite all these challenges we have proven that we can discover and characterize data that otherwise would have remained in the dark corners of the Internet. Having all this data indexed and 'triplelized' will enable scientists to access a trove of information relevant to their work in a more natural way. An important characteristic of the BCube stack is that all

  19. Dataset concerning the analytical approximation of the Ae3 temperature

    Directory of Open Access Journals (Sweden)

    B.L. Ennis

    2017-02-01

    The dataset includes the terms of the function and the values for the polynomial coefficients for major alloying elements in steel. A short description of the approximation method used to derive and validate the coefficients has also been included. For discussion and application of this model, please refer to the full length article entitled “The role of aluminium in chemical and phase segregation in a TRIP-assisted dual phase steel” 10.1016/j.actamat.2016.05.046 (Ennis et al., 2016 [1].

  20. Gene set analysis of the EADGENE chicken data-set

    DEFF Research Database (Denmark)

    Skarman, Axel; Jiang, Li; Hornshøj, Henrik

    2009-01-01

     Abstract Background: Gene set analysis is considered to be a way of improving our biological interpretation of the observed expression patterns. This paper describes different methods applied to analyse expression data from a chicken DNA microarray dataset. Results: Applying different gene set...... analyses to the chicken expression data led to different ranking of the Gene Ontology terms tested. A method for prediction of possible annotations was applied. Conclusion: Biological interpretation based on gene set analyses dependent on the statistical method used. Methods for predicting the possible...

  1. A Validation Dataset for CryoSat Sea Ice Investigators

    DEFF Research Database (Denmark)

    Julia, Gaudelli,; Baker, Steve; Haas, Christian

    Since its launch in April 2010 Cryosat has been collecting valuable sea ice data over the Arctic region. Over the same period ESA’s CryoVEx and NASA IceBridge validation campaigns have been collecting a unique set of coincident airborne measurements in the Arctic. The CryoVal-SI project has...... community. In this talk we will describe the composition of the validation dataset, summarising how it was processed and how to understand the content and format of the data. We will also explain how to access the data and the supporting documentation....

  2. Dataset of statements on policy integration of selected intergovernmental organizations

    Directory of Open Access Journals (Sweden)

    Jale Tosun

    2018-04-01

    Full Text Available This article describes data for 78 intergovernmental organizations (IGOs working on topics related to energy governance, environmental protection, and the economy. The number of IGOs covered also includes organizations active in other sectors. The point of departure for data construction was the Correlates of War dataset, from which we selected this sample of IGOs. We updated and expanded the empirical information on the IGOs selected by manual coding. Most importantly, we collected the primary law texts of the individual IGOs in order to code whether they commit themselves to environmental policy integration (EPI, climate policy integration (CPI and/or energy policy integration (EnPI.

  3. Dataset on the energy performance of atrium type hotel buildings.

    Science.gov (United States)

    Vujosevic, Milica; Krstic-Furundzic, Aleksandra

    2018-04-01

    The data presented in this article are related to the research article entitled "The Influence of Atrium on Energy Performance of Hotel Building" (Vujosevic and Krstic-Furundzic, 2017) [1], which describes the annual energy performance of atrium type hotel building in Belgrade climate conditions, with the objective to present the impact of the atrium on the hotel building's energy demands for space heating and cooling. This dataset is made publicly available to show energy performance of selected hotel design alternatives, in order to enable extended analyzes of these data for other researchers.

  4. Dataset on records of Hericium erinaceus in Slovakia.

    Science.gov (United States)

    Kunca, Vladimír; Čiliak, Marek

    2017-06-01

    The data presented in this article are related to the research article entitled "Habitat preferences of Hericium erinaceus in Slovakia" (Kunca and Čiliak, 2016) [FUNECO607] [2]. The dataset include all available and unpublished data from Slovakia, besides the records from the same tree or stem. We compiled a database of records of collections by processing data from herbaria, personal records and communication with mycological activists. Data on altitude, tree species, host tree vital status, host tree position and intensity of management of forest stands were evaluated in this study. All surveys were based on basidioma occurrence and some result from targeted searches.

  5. Dataset on records of Hericium erinaceus in Slovakia

    Directory of Open Access Journals (Sweden)

    Vladimír Kunca

    2017-06-01

    Full Text Available The data presented in this article are related to the research article entitled “Habitat preferences of Hericium erinaceus in Slovakia” (Kunca and Čiliak, 2016 [FUNECO607] [2]. The dataset include all available and unpublished data from Slovakia, besides the records from the same tree or stem. We compiled a database of records of collections by processing data from herbaria, personal records and communication with mycological activists. Data on altitude, tree species, host tree vital status, host tree position and intensity of management of forest stands were evaluated in this study. All surveys were based on basidioma occurrence and some result from targeted searches.

  6. Construction and Analysis of Long-Term Surface Temperature Dataset in Fujian Province

    Science.gov (United States)

    Li, W. E.; Wang, X. Q.; Su, H.

    2017-09-01

    Land surface temperature (LST) is a key parameter of land surface physical processes on global and regional scales, linking the heat fluxes and interactions between the ground and atmosphere. Based on MODIS 8-day LST products (MOD11A2) from the split-window algorithms, we constructed and obtained the monthly and annual LST dataset of Fujian Province from 2000 to 2015. Then, we analyzed the monthly and yearly time series LST data and further investigated the LST distribution and its evolution features. The average LST of Fujian Province reached the highest in July, while the lowest in January. The monthly and annual LST time series present a significantly periodic features (annual and interannual) from 2000 to 2015. The spatial distribution showed that the LST in North and West was lower than South and East in Fujian Province. With the rapid development and urbanization of the coastal area in Fujian Province, the LST in coastal urban region was significantly higher than that in mountainous rural region. The LST distributions might affected by the climate, topography and land cover types. The spatio-temporal distribution characteristics of LST could provide good references for the agricultural layout and environment monitoring in Fujian Province.

  7. CONSTRUCTION AND ANALYSIS OF LONG-TERM SURFACE TEMPERATURE DATASET IN FUJIAN PROVINCE

    Directory of Open Access Journals (Sweden)

    W. E. Li

    2017-09-01

    Full Text Available Land surface temperature (LST is a key parameter of land surface physical processes on global and regional scales, linking the heat fluxes and interactions between the ground and atmosphere. Based on MODIS 8-day LST products (MOD11A2 from the split-window algorithms, we constructed and obtained the monthly and annual LST dataset of Fujian Province from 2000 to 2015. Then, we analyzed the monthly and yearly time series LST data and further investigated the LST distribution and its evolution features. The average LST of Fujian Province reached the highest in July, while the lowest in January. The monthly and annual LST time series present a significantly periodic features (annual and interannual from 2000 to 2015. The spatial distribution showed that the LST in North and West was lower than South and East in Fujian Province. With the rapid development and urbanization of the coastal area in Fujian Province, the LST in coastal urban region was significantly higher than that in mountainous rural region. The LST distributions might affected by the climate, topography and land cover types. The spatio-temporal distribution characteristics of LST could provide good references for the agricultural layout and environment monitoring in Fujian Province.

  8. Temporal and spatial changes in mixed layer properties and atmospheric net heat flux in the Nordic Seas

    International Nuclear Information System (INIS)

    Smirnov, A; Alekseev, G; Korablev, A; Esau, I

    2010-01-01

    The Nordic Seas are an important area of the World Ocean where warm Atlantic waters penetrate far north forming the mild climate of Northern Europe. These waters represent the northern rim of the global thermohaline circulation. Estimates of the relationships between the net heat flux and mixed layer properties in the Nordic Seas are examined. Oceanographic data are derived from the Oceanographic Data Base (ODB) compiled in the Arctic and Antarctic Research Institute. Ocean weather ship 'Mike' (OWS) data are used to calculate radiative and turbulent components of the net heat flux. The net shortwave flux was calculated using a satellite albedo dataset and the EPA model. The net longwave flux was estimated by Southampton Oceanography Centre (SOC) method. Turbulent fluxes at the air-sea interface were calculated using the COARE 3.0 algorithm. The net heat flux was calculated by using oceanographic and meteorological data of the OWS 'Mike'. The mixed layer depth was estimated for the period since 2002 until 2009 by the 'Mike' data as well. A good correlation between these two parameters has been found. Sensible and latent heat fluxes controlled by surface air temperature/sea surface temperature gradient are the main contributors into net heat flux. Significant correlation was found between heat fluxes variations at the OWS 'Mike' location and sea ice export from the Arctic Ocean.

  9. Temporal and spatial changes in mixed layer properties and atmospheric net heat flux in the Nordic Seas

    Energy Technology Data Exchange (ETDEWEB)

    Smirnov, A; Alekseev, G [SI ' Arctic and Antarctic Research Institute' , St. Petersburg (Russian Federation); Korablev, A; Esau, I, E-mail: avsmir@aari.nw.r [Nansen Environmental and Remote Sensing Centre, Bergen (Norway)

    2010-08-15

    The Nordic Seas are an important area of the World Ocean where warm Atlantic waters penetrate far north forming the mild climate of Northern Europe. These waters represent the northern rim of the global thermohaline circulation. Estimates of the relationships between the net heat flux and mixed layer properties in the Nordic Seas are examined. Oceanographic data are derived from the Oceanographic Data Base (ODB) compiled in the Arctic and Antarctic Research Institute. Ocean weather ship 'Mike' (OWS) data are used to calculate radiative and turbulent components of the net heat flux. The net shortwave flux was calculated using a satellite albedo dataset and the EPA model. The net longwave flux was estimated by Southampton Oceanography Centre (SOC) method. Turbulent fluxes at the air-sea interface were calculated using the COARE 3.0 algorithm. The net heat flux was calculated by using oceanographic and meteorological data of the OWS 'Mike'. The mixed layer depth was estimated for the period since 2002 until 2009 by the 'Mike' data as well. A good correlation between these two parameters has been found. Sensible and latent heat fluxes controlled by surface air temperature/sea surface temperature gradient are the main contributors into net heat flux. Significant correlation was found between heat fluxes variations at the OWS 'Mike' location and sea ice export from the Arctic Ocean.

  10. Parallel Framework for Dimensionality Reduction of Large-Scale Datasets

    Directory of Open Access Journals (Sweden)

    Sai Kiranmayee Samudrala

    2015-01-01

    Full Text Available Dimensionality reduction refers to a set of mathematical techniques used to reduce complexity of the original high-dimensional data, while preserving its selected properties. Improvements in simulation strategies and experimental data collection methods are resulting in a deluge of heterogeneous and high-dimensional data, which often makes dimensionality reduction the only viable way to gain qualitative and quantitative understanding of the data. However, existing dimensionality reduction software often does not scale to datasets arising in real-life applications, which may consist of thousands of points with millions of dimensions. In this paper, we propose a parallel framework for dimensionality reduction of large-scale data. We identify key components underlying the spectral dimensionality reduction techniques, and propose their efficient parallel implementation. We show that the resulting framework can be used to process datasets consisting of millions of points when executed on a 16,000-core cluster, which is beyond the reach of currently available methods. To further demonstrate applicability of our framework we perform dimensionality reduction of 75,000 images representing morphology evolution during manufacturing of organic solar cells in order to identify how processing parameters affect morphology evolution.

  11. The Path from Large Earth Science Datasets to Information

    Science.gov (United States)

    Vicente, G. A.

    2013-12-01

    The NASA Goddard Earth Sciences Data (GES) and Information Services Center (DISC) is one of the major Science Mission Directorate (SMD) for archiving and distribution of Earth Science remote sensing data, products and services. This virtual portal provides convenient access to Atmospheric Composition and Dynamics, Hydrology, Precipitation, Ozone, and model derived datasets (generated by GSFC's Global Modeling and Assimilation Office), the North American Land Data Assimilation System (NLDAS) and the Global Land Data Assimilation System (GLDAS) data products (both generated by GSFC's Hydrological Sciences Branch). This presentation demonstrates various tools and computational technologies developed in the GES DISC to manage the huge volume of data and products acquired from various missions and programs over the years. It explores approaches to archive, document, distribute, access and analyze Earth Science data and information as well as addresses the technical and scientific issues, governance and user support problem faced by scientists in need of multi-disciplinary datasets. It also discusses data and product metrics, user distribution profiles and lessons learned through interactions with the science communities around the world. Finally it demonstrates some of the most used data and product visualization and analyses tools developed and maintained by the GES DISC.

  12. Benchmarking undedicated cloud computing providers for analysis of genomic datasets.

    Science.gov (United States)

    Yazar, Seyhan; Gooden, George E C; Mackey, David A; Hewitt, Alex W

    2014-01-01

    A major bottleneck in biological discovery is now emerging at the computational level. Cloud computing offers a dynamic means whereby small and medium-sized laboratories can rapidly adjust their computational capacity. We benchmarked two established cloud computing services, Amazon Web Services Elastic MapReduce (EMR) on Amazon EC2 instances and Google Compute Engine (GCE), using publicly available genomic datasets (E.coli CC102 strain and a Han Chinese male genome) and a standard bioinformatic pipeline on a Hadoop-based platform. Wall-clock time for complete assembly differed by 52.9% (95% CI: 27.5-78.2) for E.coli and 53.5% (95% CI: 34.4-72.6) for human genome, with GCE being more efficient than EMR. The cost of running this experiment on EMR and GCE differed significantly, with the costs on EMR being 257.3% (95% CI: 211.5-303.1) and 173.9% (95% CI: 134.6-213.1) more expensive for E.coli and human assemblies respectively. Thus, GCE was found to outperform EMR both in terms of cost and wall-clock time. Our findings confirm that cloud computing is an efficient and potentially cost-effective alternative for analysis of large genomic datasets. In addition to releasing our cost-effectiveness comparison, we present available ready-to-use scripts for establishing Hadoop instances with Ganglia monitoring on EC2 or GCE.

  13. Robust computational analysis of rRNA hypervariable tag datasets.

    Directory of Open Access Journals (Sweden)

    Maksim Sipos

    Full Text Available Next-generation DNA sequencing is increasingly being utilized to probe microbial communities, such as gastrointestinal microbiomes, where it is important to be able to quantify measures of abundance and diversity. The fragmented nature of the 16S rRNA datasets obtained, coupled with their unprecedented size, has led to the recognition that the results of such analyses are potentially contaminated by a variety of artifacts, both experimental and computational. Here we quantify how multiple alignment and clustering errors contribute to overestimates of abundance and diversity, reflected by incorrect OTU assignment, corrupted phylogenies, inaccurate species diversity estimators, and rank abundance distribution functions. We show that straightforward procedural optimizations, combining preexisting tools, are effective in handling large (10(5-10(6 16S rRNA datasets, and we describe metrics to measure the effectiveness and quality of the estimators obtained. We introduce two metrics to ascertain the quality of clustering of pyrosequenced rRNA data, and show that complete linkage clustering greatly outperforms other widely used methods.

  14. BLAST-EXPLORER helps you building datasets for phylogenetic analysis

    Directory of Open Access Journals (Sweden)

    Claverie Jean-Michel

    2010-01-01

    Full Text Available Abstract Background The right sampling of homologous sequences for phylogenetic or molecular evolution analyses is a crucial step, the quality of which can have a significant impact on the final interpretation of the study. There is no single way for constructing datasets suitable for phylogenetic analysis, because this task intimately depends on the scientific question we want to address, Moreover, database mining softwares such as BLAST which are routinely used for searching homologous sequences are not specifically optimized for this task. Results To fill this gap, we designed BLAST-Explorer, an original and friendly web-based application that combines a BLAST search with a suite of tools that allows interactive, phylogenetic-oriented exploration of the BLAST results and flexible selection of homologous sequences among the BLAST hits. Once the selection of the BLAST hits is done using BLAST-Explorer, the corresponding sequence can be imported locally for external analysis or passed to the phylogenetic tree reconstruction pipelines available on the Phylogeny.fr platform. Conclusions BLAST-Explorer provides a simple, intuitive and interactive graphical representation of the BLAST results and allows selection and retrieving of the BLAST hit sequences based a wide range of criterions. Although BLAST-Explorer primarily aims at helping the construction of sequence datasets for further phylogenetic study, it can also be used as a standard BLAST server with enriched output. BLAST-Explorer is available at http://www.phylogeny.fr

  15. Multiresolution comparison of precipitation datasets for large-scale models

    Science.gov (United States)

    Chun, K. P.; Sapriza Azuri, G.; Davison, B.; DeBeer, C. M.; Wheater, H. S.

    2014-12-01

    Gridded precipitation datasets are crucial for driving large-scale models which are related to weather forecast and climate research. However, the quality of precipitation products is usually validated individually. Comparisons between gridded precipitation products along with ground observations provide another avenue for investigating how the precipitation uncertainty would affect the performance of large-scale models. In this study, using data from a set of precipitation gauges over British Columbia and Alberta, we evaluate several widely used North America gridded products including the Canadian Gridded Precipitation Anomalies (CANGRD), the National Center for Environmental Prediction (NCEP) reanalysis, the Water and Global Change (WATCH) project, the thin plate spline smoothing algorithms (ANUSPLIN) and Canadian Precipitation Analysis (CaPA). Based on verification criteria for various temporal and spatial scales, results provide an assessment of possible applications for various precipitation datasets. For long-term climate variation studies (~100 years), CANGRD, NCEP, WATCH and ANUSPLIN have different comparative advantages in terms of their resolution and accuracy. For synoptic and mesoscale precipitation patterns, CaPA provides appealing performance of spatial coherence. In addition to the products comparison, various downscaling methods are also surveyed to explore new verification and bias-reduction methods for improving gridded precipitation outputs for large-scale models.

  16. Benchmarking Deep Learning Models on Large Healthcare Datasets.

    Science.gov (United States)

    Purushotham, Sanjay; Meng, Chuizheng; Che, Zhengping; Liu, Yan

    2018-06-04

    Deep learning models (aka Deep Neural Networks) have revolutionized many fields including computer vision, natural language processing, speech recognition, and is being increasingly used in clinical healthcare applications. However, few works exist which have benchmarked the performance of the deep learning models with respect to the state-of-the-art machine learning models and prognostic scoring systems on publicly available healthcare datasets. In this paper, we present the benchmarking results for several clinical prediction tasks such as mortality prediction, length of stay prediction, and ICD-9 code group prediction using Deep Learning models, ensemble of machine learning models (Super Learner algorithm), SAPS II and SOFA scores. We used the Medical Information Mart for Intensive Care III (MIMIC-III) (v1.4) publicly available dataset, which includes all patients admitted to an ICU at the Beth Israel Deaconess Medical Center from 2001 to 2012, for the benchmarking tasks. Our results show that deep learning models consistently outperform all the other approaches especially when the 'raw' clinical time series data is used as input features to the models. Copyright © 2018 Elsevier Inc. All rights reserved.

  17. Testing the Neutral Theory of Biodiversity with Human Microbiome Datasets.

    Science.gov (United States)

    Li, Lianwei; Ma, Zhanshan Sam

    2016-08-16

    The human microbiome project (HMP) has made it possible to test important ecological theories for arguably the most important ecosystem to human health-the human microbiome. Existing limited number of studies have reported conflicting evidence in the case of the neutral theory; the present study aims to comprehensively test the neutral theory with extensive HMP datasets covering all five major body sites inhabited by the human microbiome. Utilizing 7437 datasets of bacterial community samples, we discovered that only 49 communities (less than 1%) satisfied the neutral theory, and concluded that human microbial communities are not neutral in general. The 49 positive cases, although only a tiny minority, do demonstrate the existence of neutral processes. We realize that the traditional doctrine of microbial biogeography "Everything is everywhere, but the environment selects" first proposed by Baas-Becking resolves the apparent contradiction. The first part of Baas-Becking doctrine states that microbes are not dispersal-limited and therefore are neutral prone, and the second part reiterates that the freely dispersed microbes must endure selection by the environment. Therefore, in most cases, it is the host environment that ultimately shapes the community assembly and tip the human microbiome to niche regime.

  18. Overview of the CERES Edition-4 Multilayer Cloud Property Datasets

    Science.gov (United States)

    Chang, F. L.; Minnis, P.; Sun-Mack, S.; Chen, Y.; Smith, R. A.; Brown, R. R.

    2014-12-01

    Knowledge of the cloud vertical distribution is important for understanding the role of clouds on earth's radiation budget and climate change. Since high-level cirrus clouds with low emission temperatures and small optical depths can provide a positive feedback to a climate system and low-level stratus clouds with high emission temperatures and large optical depths can provide a negative feedback effect, the retrieval of multilayer cloud properties using satellite observations, like Terra and Aqua MODIS, is critically important for a variety of cloud and climate applications. For the objective of the Clouds and the Earth's Radiant Energy System (CERES), new algorithms have been developed using Terra and Aqua MODIS data to allow separate retrievals of cirrus and stratus cloud properties when the two dominant cloud types are simultaneously present in a multilayer system. In this paper, we will present an overview of the new CERES Edition-4 multilayer cloud property datasets derived from Terra as well as Aqua. Assessment of the new CERES multilayer cloud datasets will include high-level cirrus and low-level stratus cloud heights, pressures, and temperatures as well as their optical depths, emissivities, and microphysical properties.

  19. Predicting weather regime transitions in Northern Hemisphere datasets

    Energy Technology Data Exchange (ETDEWEB)

    Kondrashov, D. [University of California, Department of Atmospheric and Oceanic Sciences and Institute of Geophysics and Planetary Physics, Los Angeles, CA (United States); Shen, J. [UCLA, Department of Statistics, Los Angeles, CA (United States); Berk, R. [UCLA, Department of Statistics, Los Angeles, CA (United States); University of Pennsylvania, Department of Criminology, Philadelphia, PA (United States); D' Andrea, F.; Ghil, M. [Ecole Normale Superieure, Departement Terre-Atmosphere-Ocean and Laboratoire de Meteorologie Dynamique (CNRS and IPSL), Paris Cedex 05 (France)

    2007-10-15

    A statistical learning method called random forests is applied to the prediction of transitions between weather regimes of wintertime Northern Hemisphere (NH) atmospheric low-frequency variability. A dataset composed of 55 winters of NH 700-mb geopotential height anomalies is used in the present study. A mixture model finds that the three Gaussian components that were statistically significant in earlier work are robust; they are the Pacific-North American (PNA) regime, its approximate reverse (the reverse PNA, or RNA), and the blocked phase of the North Atlantic Oscillation (BNAO). The most significant and robust transitions in the Markov chain generated by these regimes are PNA {yields} BNAO, PNA {yields} RNA and BNAO {yields} PNA. The break of a regime and subsequent onset of another one is forecast for these three transitions. Taking the relative costs of false positives and false negatives into account, the random-forests method shows useful forecasting skill. The calculations are carried out in the phase space spanned by a few leading empirical orthogonal functions of dataset variability. Plots of estimated response functions to a given predictor confirm the crucial influence of the exit angle on a preferred transition path. This result points to the dynamic origin of the transitions. (orig.)

  20. Digital Astronaut Photography: A Discovery Dataset for Archaeology

    Science.gov (United States)

    Stefanov, William L.

    2010-01-01

    Astronaut photography acquired from the International Space Station (ISS) using commercial off-the-shelf cameras offers a freely-accessible source for high to very high resolution (4-20 m/pixel) visible-wavelength digital data of Earth. Since ISS Expedition 1 in 2000, over 373,000 images of the Earth-Moon system (including land surface, ocean, atmospheric, and lunar images) have been added to the Gateway to Astronaut Photography of Earth online database (http://eol.jsc.nasa.gov ). Handheld astronaut photographs vary in look angle, time of acquisition, solar illumination, and spatial resolution. These attributes of digital astronaut photography result from a unique combination of ISS orbital dynamics, mission operations, camera systems, and the individual skills of the astronaut. The variable nature of astronaut photography makes the dataset uniquely useful for archaeological applications in comparison with more traditional nadir-viewing multispectral datasets acquired from unmanned orbital platforms. For example, surface features such as trenches, walls, ruins, urban patterns, and vegetation clearing and regrowth patterns may be accentuated by low sun angles and oblique viewing conditions (Fig. 1). High spatial resolution digital astronaut photographs can also be used with sophisticated land cover classification and spatial analysis approaches like Object Based Image Analysis, increasing the potential for use in archaeological characterization of landscapes and specific sites.

  1. ISC-EHB: Reconstruction of a robust earthquake dataset

    Science.gov (United States)

    Weston, J.; Engdahl, E. R.; Harris, J.; Di Giacomo, D.; Storchak, D. A.

    2018-04-01

    The EHB Bulletin of hypocentres and associated travel-time residuals was originally developed with procedures described by Engdahl, Van der Hilst and Buland (1998) and currently ends in 2008. It is a widely used seismological dataset, which is now expanded and reconstructed, partly by exploiting updated procedures at the International Seismological Centre (ISC), to produce the ISC-EHB. The reconstruction begins in the modern period (2000-2013) to which new and more rigorous procedures for event selection, data preparation, processing, and relocation are applied. The selection criteria minimise the location bias produced by unmodelled 3D Earth structure, resulting in events that are relatively well located in any given region. Depths of the selected events are significantly improved by a more comprehensive review of near station and secondary phase travel-time residuals based on ISC data, especially for the depth phases pP, pwP and sP, as well as by a rigorous review of the event depths in subduction zone cross sections. The resulting cross sections and associated maps are shown to provide details of seismicity in subduction zones in much greater detail than previously achievable. The new ISC-EHB dataset will be especially useful for global seismicity studies and high-frequency regional and global tomographic inversions.

  2. Benchmarking undedicated cloud computing providers for analysis of genomic datasets.

    Directory of Open Access Journals (Sweden)

    Seyhan Yazar

    Full Text Available A major bottleneck in biological discovery is now emerging at the computational level. Cloud computing offers a dynamic means whereby small and medium-sized laboratories can rapidly adjust their computational capacity. We benchmarked two established cloud computing services, Amazon Web Services Elastic MapReduce (EMR on Amazon EC2 instances and Google Compute Engine (GCE, using publicly available genomic datasets (E.coli CC102 strain and a Han Chinese male genome and a standard bioinformatic pipeline on a Hadoop-based platform. Wall-clock time for complete assembly differed by 52.9% (95% CI: 27.5-78.2 for E.coli and 53.5% (95% CI: 34.4-72.6 for human genome, with GCE being more efficient than EMR. The cost of running this experiment on EMR and GCE differed significantly, with the costs on EMR being 257.3% (95% CI: 211.5-303.1 and 173.9% (95% CI: 134.6-213.1 more expensive for E.coli and human assemblies respectively. Thus, GCE was found to outperform EMR both in terms of cost and wall-clock time. Our findings confirm that cloud computing is an efficient and potentially cost-effective alternative for analysis of large genomic datasets. In addition to releasing our cost-effectiveness comparison, we present available ready-to-use scripts for establishing Hadoop instances with Ganglia monitoring on EC2 or GCE.

  3. Condensing Massive Satellite Datasets For Rapid Interactive Analysis

    Science.gov (United States)

    Grant, G.; Gallaher, D. W.; Lv, Q.; Campbell, G. G.; Fowler, C.; LIU, Q.; Chen, C.; Klucik, R.; McAllister, R. A.

    2015-12-01

    Our goal is to enable users to interactively analyze massive satellite datasets, identifying anomalous data or values that fall outside of thresholds. To achieve this, the project seeks to create a derived database containing only the most relevant information, accelerating the analysis process. The database is designed to be an ancillary tool for the researcher, not an archival database to replace the original data. This approach is aimed at improving performance by reducing the overall size by way of condensing the data. The primary challenges of the project include: - The nature of the research question(s) may not be known ahead of time. - The thresholds for determining anomalies may be uncertain. - Problems associated with processing cloudy, missing, or noisy satellite imagery. - The contents and method of creation of the condensed dataset must be easily explainable to users. The architecture of the database will reorganize spatially-oriented satellite imagery into temporally-oriented columns of data (a.k.a., "data rods") to facilitate time-series analysis. The database itself is an open-source parallel database, designed to make full use of clustered server technologies. A demonstration of the system capabilities will be shown. Applications for this technology include quick-look views of the data, as well as the potential for on-board satellite processing of essential information, with the goal of reducing data latency.

  4. Utilizing the Antarctic Master Directory to find orphan datasets

    Science.gov (United States)

    Bonczkowski, J.; Carbotte, S. M.; Arko, R. A.; Grebas, S. K.

    2011-12-01

    While most Antarctic data are housed at an established disciplinary-specific data repository, there are data types for which no suitable repository exists. In some cases, these "orphan" data, without an appropriate national archive, are served from local servers by the principal investigators who produced the data. There are many pitfalls with data served privately, including the frequent lack of adequate documentation to ensure the data can be understood by others for re-use and the impermanence of personal web sites. For example, if an investigator leaves an institution and the data moves, the link published is no longer accessible. To ensure continued availability of data, submission to long-term national data repositories is needed. As stated in the National Science Foundation Office of Polar Programs (NSF/OPP) Guidelines and Award Conditions for Scientific Data, investigators are obligated to submit their data for curation and long-term preservation; this includes the registration of a dataset description into the Antarctic Master Directory (AMD), http://gcmd.nasa.gov/Data/portals/amd/. The AMD is a Web-based, searchable directory of thousands of dataset descriptions, known as DIF records, submitted by scientists from over 20 countries. It serves as a node of the International Directory Network/Global Change Master Directory (IDN/GCMD). The US Antarctic Program Data Coordination Center (USAP-DCC), http://www.usap-data.org/, funded through NSF/OPP, was established in 2007 to help streamline the process of data submission and DIF record creation. When data does not quite fit within any existing disciplinary repository, it can be registered within the USAP-DCC as the fallback data repository. Within the scope of the USAP-DCC we undertook the challenge of discovering and "rescuing" orphan datasets currently registered within the AMD. In order to find which DIF records led to data served privately, all records relating to US data within the AMD were parsed. After

  5. Flavour mixings in flux compactifications

    International Nuclear Information System (INIS)

    Buchmuller, Wilfried; Schweizer, Julian

    2017-01-01

    A multiplicity of quark-lepton families can naturally arise as zero-modes in flux compactifications. The flavour structure of quark and lepton mass matrices is then determined by the wave function profiles of the zero-modes. We consider a supersymmetric SO(10) x U(1) model in six dimensions compactified on the orbifold T 2 =Z 2 with Abelian magnetic flux. A bulk 16-plet charged under the U(1) provides the quark-lepton generations whereas two uncharged 10-plets yield two Higgs doublets. Bulk anomaly cancellation requires the presence of additional 16- and 10-plets. The corresponding zero-modes form vectorlike split multiplets that are needed to obtain a successful flavour phenomenology. We analyze the pattern of flavour mixings for the two heaviest families of the Standard Model and discuss possible generalizations to three and more generations.

  6. Superconducting flux flow digital circuits

    International Nuclear Information System (INIS)

    Martens, J.S.; Zipperian, T.E.; Hietala, V.M.; Ginley, D.S.; Tigges, C.P.; Phillips, J.M.; Siegal, M.P.

    1993-01-01

    The authors have developed a family of digital logic circuits based on superconducting flux flow transistors that show high speed, reasonable signal levels, large fan-out, and large noise margins. The circuits are made from high-temperature superconductors (HTS) and have been shown to operate at over 90 K. NOR gates have been demonstrated with fan-outs of more than 5 and fully loaded switching times less than a fixture-limited 50 ps. Ring-oscillator data suggest inverter delay times of about 40ps when using a 3-μm linewidths. Simple flip-flops have also been demonstrated showing large noise margins, response times of less than 30 ps, and static power dissipation on the order of 30 nW. Among other uses, this logic family is appropriate as an interface between logic families such as single flux quantum and conventional semiconductor logic

  7. Heisenberg groups and noncommutative fluxes

    International Nuclear Information System (INIS)

    Freed, Daniel S.; Moore, Gregory W.; Segal, Graeme

    2007-01-01

    We develop a group-theoretical approach to the formulation of generalized abelian gauge theories, such as those appearing in string theory and M-theory. We explore several applications of this approach. First, we show that there is an uncertainty relation which obstructs simultaneous measurement of electric and magnetic flux when torsion fluxes are included. Next, we show how to define the Hilbert space of a self-dual field. The Hilbert space is Z 2 -graded and we show that, in general, self-dual theories (including the RR fields of string theory) have fermionic sectors. We indicate how rational conformal field theories associated to the two-dimensional Gaussian model generalize to (4k+2)-dimensional conformal field theories. When our ideas are applied to the RR fields of string theory we learn that it is impossible to measure the K-theory class of a RR field. Only the reduction modulo torsion can be measured

  8. Neutron flux enhancement at LASREF

    International Nuclear Information System (INIS)

    Sommer, W.F.; Ferguson, P.D.; Wechsler, M.S.

    1992-01-01

    The accelerator at the Los Alamos Meson Physiscs Facility produces a 1 mA beam of protons at an energy of 800 MeV. Since 1985, the Los Alamos Spallation Radiation Effects Facility (LASREF) has made use of the neutron flux that is generated as the incident protons interact with the targets and a copper beam stop. A variety of basic and applied experiments in radiation damage and radiation effects have been completed. Recent studies indicate that the flux at LASREF can be increased by at least a factor of 10 from the present level of about 5 E + 17 m -2 s -1 . This requires changing the beam stop material from Cu to W and optimizing the geometry of the beam-target interaction region. These studies are motivated by the need for a large volume, high energy, and high intensity neutron source in the development of materials for advanced energy concepts such as fusion reactors. (orig.)

  9. Flux through a Markov chain

    International Nuclear Information System (INIS)

    Floriani, Elena; Lima, Ricardo; Ourrad, Ouerdia; Spinelli, Lionel

    2016-01-01

    Highlights: • The flux through a Markov chain of a conserved quantity (mass) is studied. • Mass is supplied by an external source and ends in the absorbing states of the chain. • Meaningful for modeling open systems whose dynamics has a Markov property. • The analytical expression of mass distribution is given for a constant source. • The expression of mass distribution is given for periodic or random sources. - Abstract: In this paper we study the flux through a finite Markov chain of a quantity, that we will call mass, which moves through the states of the chain according to the Markov transition probabilities. Mass is supplied by an external source and accumulates in the absorbing states of the chain. We believe that studying how this conserved quantity evolves through the transient (non-absorbing) states of the chain could be useful for the modelization of open systems whose dynamics has a Markov property.

  10. Absolute flux scale for radioastronomy

    International Nuclear Information System (INIS)

    Ivanov, V.P.; Stankevich, K.S.

    1986-01-01

    The authors propose and provide support for a new absolute flux scale for radio astronomy, which is not encumbered with the inadequacies of the previous scales. In constructing it the method of relative spectra was used (a powerful tool for choosing reference spectra). A review is given of previous flux scales. The authors compare the AIS scale with the scale they propose. Both scales are based on absolute measurements by the ''artificial moon'' method, and they are practically coincident in the range from 0.96 to 6 GHz. At frequencies above 6 GHz, 0.96 GHz, the AIS scale is overestimated because of incorrect extrapolation of the spectra of the primary and secondary standards. The major results which have emerged from this review of absolute scales in radio astronomy are summarized

  11. Rapid reconnection of flux lines

    International Nuclear Information System (INIS)

    Samain, A.

    1982-01-01

    The rapid reconnection of flux lines in an incompressible fluid through a singular layer of the current density is discussed. It is shown that the liberated magnetic energy must partially appear in the form of plasma kinetic energy. A laminar structure of the flow is possible, but Alfven velocity must be achieved in eddies of growing size at the ends of the layer. The gross structure of the flow and the magnetic configuration may be obtained from variational principles. (author)

  12. NERIES: Seismic Data Gateways and User Composed Datasets Metadata Management

    Science.gov (United States)

    Spinuso, Alessandro; Trani, Luca; Kamb, Linus; Frobert, Laurent

    2010-05-01

    One of the NERIES EC project main objectives is to establish and improve the networking of seismic waveform data exchange and access among four main data centers in Europe: INGV, GFZ, ORFEUS and IPGP. Besides the implementation of the data backbone, several investigations and developments have been conducted in order to offer to the users the data available from this network, either programmatically or interactively. One of the challenges is to understand how to enable users` activities such as discovering, aggregating, describing and sharing datasets to obtain a decrease in the replication of similar data queries towards the network, exempting the data centers to guess and create useful pre-packed products. We`ve started to transfer this task more and more towards the users community, where the users` composed data products could be extensively re-used. The main link to the data is represented by a centralized webservice (SeismoLink) acting like a single access point to the whole data network. Users can download either waveform data or seismic station inventories directly from their own software routines by connecting to this webservice, which routes the request to the data centers. The provenance of the data is maintained and transferred to the users in the form of URIs, that identify the dataset and implicitly refer to the data provider. SeismoLink, combined with other webservices (eg EMSC-QuakeML earthquakes catalog service), is used from a community gateway such as the NERIES web portal (http://www.seismicportal.eu). Here the user interacts with a map based portlet which allows the dynamic composition of a data product, binding seismic event`s parameters with a set of seismic stations. The requested data is collected by the back-end processes of the portal, preserved and offered to the user in a personal data cart, where metadata can be generated interactively on-demand. The metadata, expressed in RDF, can also be remotely ingested. They offer rating

  13. Accuracy assessment of seven global land cover datasets over China

    Science.gov (United States)

    Yang, Yongke; Xiao, Pengfeng; Feng, Xuezhi; Li, Haixing

    2017-03-01

    Land cover (LC) is the vital foundation to Earth science. Up to now, several global LC datasets have arisen with efforts of many scientific communities. To provide guidelines for data usage over China, nine LC maps from seven global LC datasets (IGBP DISCover, UMD, GLC, MCD12Q1, GLCNMO, CCI-LC, and GlobeLand30) were evaluated in this study. First, we compared their similarities and discrepancies in both area and spatial patterns, and analysed their inherent relations to data sources and classification schemes and methods. Next, five sets of validation sample units (VSUs) were collected to calculate their accuracy quantitatively. Further, we built a spatial analysis model and depicted their spatial variation in accuracy based on the five sets of VSUs. The results show that, there are evident discrepancies among these LC maps in both area and spatial patterns. For LC maps produced by different institutes, GLC 2000 and CCI-LC 2000 have the highest overall spatial agreement (53.8%). For LC maps produced by same institutes, overall spatial agreement of CCI-LC 2000 and 2010, and MCD12Q1 2001 and 2010 reach up to 99.8% and 73.2%, respectively; while more efforts are still needed if we hope to use these LC maps as time series data for model inputting, since both CCI-LC and MCD12Q1 fail to represent the rapid changing trend of several key LC classes in the early 21st century, in particular urban and built-up, snow and ice, water bodies, and permanent wetlands. With the highest spatial resolution, the overall accuracy of GlobeLand30 2010 is 82.39%. For the other six LC datasets with coarse resolution, CCI-LC 2010/2000 has the highest overall accuracy, and following are MCD12Q1 2010/2001, GLC 2000, GLCNMO 2008, IGBP DISCover, and UMD in turn. Beside that all maps exhibit high accuracy in homogeneous regions; local accuracies in other regions are quite different, particularly in Farming-Pastoral Zone of North China, mountains in Northeast China, and Southeast Hills. Special

  14. Neutron flux control systems validation

    International Nuclear Information System (INIS)

    Hascik, R.

    2003-01-01

    In nuclear installations main requirement is to obtain corresponding nuclear safety in all operation conditions. From the nuclear safety point of view is commissioning and start-up after reactor refuelling appropriate period for safety systems verification. In this paper, methodology, performance and results of neutron flux measurements systems validation is presented. Standard neutron flux measuring chains incorporated into the reactor protection and control system are used. Standard neutron flux measuring chain contains detector, preamplifier, wiring to data acquisition unit, data acquisition unit, wiring to display at control room and display at control room. During reactor outage only data acquisition unit and wiring and displaying at reactor control room is verified. It is impossible to verify detector, preamplifier and wiring to data acquisition recording unit during reactor refuelling according to low power. Adjustment and accurate functionality of these chains is confirmed by start-up rate (SUR) measurement during start-up tests after refuelling of the reactors. This measurement has direct impact to nuclear safety and increase operational nuclear safety level. Briefly description of each measuring system is given. Results are illustrated on measurements performed at Bohunice NPP during reactor start-up tests. Main failures and their elimination are described (Authors)

  15. Surface fluxes in heterogeneous landscape

    Energy Technology Data Exchange (ETDEWEB)

    Bay Hasager, C

    1997-01-01

    The surface fluxes in homogeneous landscapes are calculated by similarity scaling principles. The methodology is well establish. In heterogeneous landscapes with spatial changes in the micro scale range, i e from 100 m to 10 km, advective effects are significant. The present work focus on these effects in an agricultural countryside typical for the midlatitudes. Meteorological and satellite data from a highly heterogeneous landscape in the Rhine Valley, Germany was collected in the large-scale field experiment TRACT (Transport of pollutants over complex terrain) in 1992. Classified satellite images, Landsat TM and ERS SAR, are used as basis for roughness maps. The roughnesses were measured at meteorological masts in the various cover classes and assigned pixel by pixel to the images. The roughness maps are aggregated, i e spatially averaged, into so-called effective roughness lengths. This calculation is performed by a micro scale aggregation model. The model solves the linearized atmospheric flow equations by a numerical (Fast Fourier Transform) method. This model also calculate maps of friction velocity and momentum flux pixel wise in heterogeneous landscapes. It is indicated how the aggregation methodology can be used to calculate the heat fluxes based on the relevant satellite data i e temperature and soil moisture information. (au) 10 tabs., 49 ills., 223 refs.

  16. Generalized drift-flux correlation

    International Nuclear Information System (INIS)

    Takeuchi, K.; Young, M.Y.; Hochreiter, L.E.

    1991-01-01

    A one-dimensional drift-flux model with five conservation equations is frequently employed in major computer codes, such as TRAC-PD2, and in simulator codes. In this method, the relative velocity between liquid and vapor phases, or slip ratio, is given by correlations, rather than by direct solution of the phasic momentum equations, as in the case of the two-fluid model used in TRAC-PF1. The correlations for churn-turbulent bubbly flow and slug flow regimes were given in terms of drift velocities by Zuber and Findlay. For the annular flow regime, the drift velocity correlations were developed by Ishii et al., using interphasic force balances. Another approach is to define the drift velocity so that flooding and liquid hold-up conditions are properly simulated, as reported here. The generalized correlation is used to reanalyze the MB-2 test data for two-phase flow in a large-diameter pipe. The results are applied to the generalized drift flux velocity, whose relationship to the other correlations is discussed. Finally, the generalized drift flux correlation is implemented in TRAC-PD2. Flow reversal from countercurrent to cocurrent flow is computed in small-diameter U-shaped tubes and is compared with the flooding curve

  17. Gridded 5km GHCN-Daily Temperature and Precipitation Dataset, Version 1

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The Gridded 5km GHCN-Daily Temperature and Precipitation Dataset (nClimGrid) consists of four climate variables derived from the GHCN-D dataset: maximum temperature,...

  18. Active Semisupervised Clustering Algorithm with Label Propagation for Imbalanced and Multidensity Datasets

    Directory of Open Access Journals (Sweden)

    Mingwei Leng

    2013-01-01

    Full Text Available The accuracy of most of the existing semisupervised clustering algorithms based on small size of labeled dataset is low when dealing with multidensity and imbalanced datasets, and labeling data is quite expensive and time consuming in many real-world applications. This paper focuses on active data selection and semisupervised clustering algorithm in multidensity and imbalanced datasets and proposes an active semisupervised clustering algorithm. The proposed algorithm uses an active mechanism for data selection to minimize the amount of labeled data, and it utilizes multithreshold to expand labeled datasets on multidensity and imbalanced datasets. Three standard datasets and one synthetic dataset are used to demonstrate the proposed algorithm, and the experimental results show that the proposed semisupervised clustering algorithm has a higher accuracy and a more stable performance in comparison to other clustering and semisupervised clustering algorithms, especially when the datasets are multidensity and imbalanced.

  19. Dataset for Probabilistic estimation of residential air exchange rates for population-based exposure modeling

    Data.gov (United States)

    U.S. Environmental Protection Agency — This dataset provides the city-specific air exchange rate measurements, modeled, literature-based as well as housing characteristics. This dataset is associated with...

  20. An Affinity Propagation Clustering Algorithm for Mixed Numeric and Categorical Datasets

    Directory of Open Access Journals (Sweden)

    Kang Zhang

    2014-01-01

    Full Text Available Clustering has been widely used in different fields of science, technology, social science, and so forth. In real world, numeric as well as categorical features are usually used to describe the data objects. Accordingly, many clustering methods can process datasets that are either numeric or categorical. Recently, algorithms that can handle the mixed data clustering problems have been developed. Affinity propagation (AP algorithm is an exemplar-based clustering method which has demonstrated good performance on a wide variety of datasets. However, it has limitations on processing mixed datasets. In this paper, we propose a novel similarity measure for mixed type datasets and an adaptive AP clustering algorithm is proposed to cluster the mixed datasets. Several real world datasets are studied to evaluate the performance of the proposed algorithm. Comparisons with other clustering algorithms demonstrate that the proposed method works well not only on mixed datasets but also on pure numeric and categorical datasets.

  1. Ecohydrological Index, Native Fish, and Climate Trends and Relationships in the Kansas River Basin_dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — The dataset is an excel file that contain data for the figures in the manuscript. This dataset is associated with the following publication: Sinnathamby, S., K....

  2. Global Human Built-up And Settlement Extent (HBASE) Dataset From Landsat

    Data.gov (United States)

    National Aeronautics and Space Administration — The Global Human Built-up And Settlement Extent (HBASE) Dataset from Landsat is a global map of HBASE derived from the Global Land Survey (GLS) Landsat dataset for...

  3. Quantitative intracellular flux modeling and applications in biotherapeutic development and production using CHO cell cultures.

    Science.gov (United States)

    Huang, Zhuangrong; Lee, Dong-Yup; Yoon, Seongkyu

    2017-12-01

    Chinese hamster ovary (CHO) cells have been widely used for producing many recombinant therapeutic proteins. Constraint-based modeling, such as flux balance analysis (FBA) and metabolic flux analysis (MFA), has been developing rapidly for the quantification of intracellular metabolic flux distribution at a systematic level. Such methods would produce detailed maps of flows through metabolic networks, which contribute significantly to better understanding of metabolism in cells. Although these approaches have been extensively established in microbial systems, their application to mammalian cells is sparse. This review brings together the recent development of constraint-based models and their applications in CHO cells. The further development of constraint-based modeling approaches driven by multi-omics datasets is discussed, and a framework of potential modeling application in cell culture engineering is proposed. Improved cell culture system understanding will enable robust developments in cell line and bioprocess engineering thus accelerating consistent process quality control in biopharmaceutical manufacturing. © 2017 Wiley Periodicals, Inc.

  4. Periodicities observed on solar flux index (F10.7) during geomagnetic disturbances

    Science.gov (United States)

    Adhikari, B.; Narayan, C.; Chhatkuli, D. N.

    2017-12-01

    Solar activities change within the period of 11 years. Sometimes the greatest event occurs in the period of solar maxima and the lowest activity occurs in the period of solar minimum. During the time period of solar activity sunspots number will vary. A 10.7 cm solar flux measurement is a determination of the strength of solar radio emission. The solar flux index is more often used for the prediction and monitoring of the solar activity. This study mainly focused on the variation on solar flux index and amount of electromagnetic wave in the atmosphere. Both seasonal and yearly variation on solar F10.7 index. We also analyzed the dataset obatained from riometer.Both instruments show seasonal and yearly variations. We also observed the solar cycle dependence on solar flux index and found a strong dependence on solar activity. Results also show that solar intensities higher during the rising phase of solar cycle. We also observed periodicities on solar flux index using wavelet analysis. Through this analysis, it was found that the power intensities of solar flux index show a high spectral variability.

  5. Quantifying the Terrestrial Surface Energy Fluxes Using Remotely-Sensed Satellite Data

    Science.gov (United States)

    Siemann, Amanda Lynn

    The dynamics of the energy fluxes between the land surface and the atmosphere drive local and regional climate and are paramount to understand the past, present, and future changes in climate. Although global reanalysis datasets, land surface models (LSMs), and climate models estimate these fluxes by simulating the physical processes involved, they merely simulate our current understanding of these processes. Global estimates of the terrestrial, surface energy fluxes based on observations allow us to capture the dynamics of the full climate system. Remotely-sensed satellite data is the source of observations of the land surface which provide the widest spatial coverage. Although net radiation and latent heat flux global, terrestrial, surface estimates based on remotely-sensed satellite data have progressed, comparable sensible heat data products and ground heat flux products have not progressed at this scale. Our primary objective is quantifying and understanding the terrestrial energy fluxes at the Earth's surface using remotely-sensed satellite data with consistent development among all energy budget components [through the land surface temperature (LST) and input meteorology], including validation of these products against in-situ data, uncertainty assessments, and long-term trend analysis. The turbulent fluxes are constrained by the available energy using the Bowen ratio of the un-constrained products to ensure energy budget closure. All final products are within uncertainty ranges of literature values, globally. When validated against the in-situ estimates, the sensible heat flux estimates using the CFSR air temperature and constrained with the products using the MODIS albedo produce estimates closest to the FLUXNET in-situ observations. Poor performance over South America is consistent with the largest uncertainties in the energy budget. From 1984-2007, the longwave upward flux increase due to the LST increase drives the net radiation decrease, and the

  6. An Automatic Matcher and Linker for Transportation Datasets

    Directory of Open Access Journals (Sweden)

    Ali Masri

    2017-01-01

    Full Text Available Multimodality requires the integration of heterogeneous transportation data to construct a broad view of the transportation network. Many new transportation services are emerging while being isolated from previously-existing networks. This leads them to publish their data sources to the web, according to linked data principles, in order to gain visibility. Our interest is to use these data to construct an extended transportation network that links these new services to existing ones. The main problems we tackle in this article fall in the categories of automatic schema matching and data interlinking. We propose an approach that uses web services as mediators to help in automatically detecting geospatial properties and mapping them between two different schemas. On the other hand, we propose a new interlinking approach that enables the user to define rich semantic links between datasets in a flexible and customizable way.

  7. [Parallel virtual reality visualization of extreme large medical datasets].

    Science.gov (United States)

    Tang, Min

    2010-04-01

    On the basis of a brief description of grid computing, the essence and critical techniques of parallel visualization of extreme large medical datasets are discussed in connection with Intranet and common-configuration computers of hospitals. In this paper are introduced several kernel techniques, including the hardware structure, software framework, load balance and virtual reality visualization. The Maximum Intensity Projection algorithm is realized in parallel using common PC cluster. In virtual reality world, three-dimensional models can be rotated, zoomed, translated and cut interactively and conveniently through the control panel built on virtual reality modeling language (VRML). Experimental results demonstrate that this method provides promising and real-time results for playing the role in of a good assistant in making clinical diagnosis.

  8. The wildland-urban interface raster dataset of Catalonia.

    Science.gov (United States)

    Alcasena, Fermín J; Evers, Cody R; Vega-Garcia, Cristina

    2018-04-01

    We provide the wildland urban interface (WUI) map of the autonomous community of Catalonia (Northeastern Spain). The map encompasses an area of some 3.21 million ha and is presented as a 150-m resolution raster dataset. Individual housing location, structure density and vegetation cover data were used to spatially assess in detail the interface, intermix and dispersed rural WUI communities with a geographical information system. Most WUI areas concentrate in the coastal belt where suburban sprawl has occurred nearby or within unmanaged forests. This geospatial information data provides an approximation of residential housing potential for loss given a wildfire, and represents a valuable contribution to assist landscape and urban planning in the region.

  9. xarray: N-D labeled Arrays and Datasets in Python

    Directory of Open Access Journals (Sweden)

    Stephan Hoyer

    2017-04-01

    Full Text Available xarray is an open source project and Python package that provides a toolkit and data structures for N-dimensional labeled arrays. Our approach combines an application programing interface (API inspired by pandas with the Common Data Model for self-described scientific data. Key features of the xarray package include label-based indexing and arithmetic, interoperability with the core scientific Python packages (e.g., pandas, NumPy, Matplotlib, out-of-core computation on datasets that don’t fit into memory, a wide range of serialization and input/output (I/O options, and advanced multi-dimensional data manipulation tools such as group-by and resampling. xarray, as a data model and analytics toolkit, has been widely adopted in the geoscience community but is also used more broadly for multi-dimensional data analysis in physics, machine learning and finance.

  10. The wildland-urban interface raster dataset of Catalonia

    Directory of Open Access Journals (Sweden)

    Fermín J. Alcasena

    2018-04-01

    Full Text Available We provide the wildland urban interface (WUI map of the autonomous community of Catalonia (Northeastern Spain. The map encompasses an area of some 3.21 million ha and is presented as a 150-m resolution raster dataset. Individual housing location, structure density and vegetation cover data were used to spatially assess in detail the interface, intermix and dispersed rural WUI communities with a geographical information system. Most WUI areas concentrate in the coastal belt where suburban sprawl has occurred nearby or within unmanaged forests. This geospatial information data provides an approximation of residential housing potential for loss given a wildfire, and represents a valuable contribution to assist landscape and urban planning in the region. Keywords: Wildland-urban interface, Wildfire risk, Urban planning, Human communities, Catalonia

  11. Reconstructing flaw image using dataset of full matrix capture technique

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Tae Hun; Kim, Yong Sik; Lee, Jeong Seok [KHNP Central Research Institute, Daejeon (Korea, Republic of)

    2017-02-15

    A conventional phased array ultrasonic system offers the ability to steer an ultrasonic beam by applying independent time delays of individual elements in the array and produce an ultrasonic image. In contrast, full matrix capture (FMC) is a data acquisition process that collects a complete matrix of A-scans from every possible independent transmit-receive combination in a phased array transducer and makes it possible to reconstruct various images that cannot be produced by conventional phased array with the post processing as well as images equivalent to a conventional phased array image. In this paper, a basic algorithm based on the LLL mode total focusing method (TFM) that can image crack type flaws is described. And this technique was applied to reconstruct flaw images from the FMC dataset obtained from the experiments and ultrasonic simulation.

  12. Survey dataset on occupational hazards on construction sites

    Directory of Open Access Journals (Sweden)

    Patience F. Tunji-Olayeni

    2018-06-01

    Full Text Available The construction site provides an unfriendly working conditions, exposing workers to one of the harshest environments at a workplace. In this dataset, a structured questionnaire was design directed to thirty-five (35 craftsmen selected through a purposive sampling technique on various construction sites in one of the most populous cities in sub-Saharan Africa. The set of descriptive statistics is presented with tables, stacked bar chats and pie charts. Common occupational health conditions affecting the cardiovascular, respiratory and musculoskeletal systems of craftsmen on construction sites were identified. The effects of occupational health hazards on craftsmen and on construction project performance can be determined when the data is analyzed. Moreover, contractors’ commitment to occupational health and safety (OHS can be obtained from the analysis of the survey data. Keywords: Accidents, Construction industry, Craftsmen, Health, Occupational hazards

  13. Feedback control in deep drawing based on experimental datasets

    Science.gov (United States)

    Fischer, P.; Heingärtner, J.; Aichholzer, W.; Hortig, D.; Hora, P.

    2017-09-01

    In large-scale production of deep drawing parts, like in automotive industry, the effects of scattering material properties as well as warming of the tools have a significant impact on the drawing result. In the scope of the work, an approach is presented to minimize the influence of these effects on part quality by optically measuring the draw-in of each part and adjusting the settings of the press to keep the strain distribution, which is represented by the draw-in, inside a certain limit. For the design of the control algorithm, a design of experiments for in-line tests is used to quantify the influence of the blank holder force as well as the force distribution on the draw-in. The results of this experimental dataset are used to model the process behavior. Based on this model, a feedback control loop is designed. Finally, the performance of the control algorithm is validated in the production line.

  14. Orthology detection combining clustering and synteny for very large datasets.

    Science.gov (United States)

    Lechner, Marcus; Hernandez-Rosales, Maribel; Doerr, Daniel; Wieseke, Nicolas; Thévenin, Annelyse; Stoye, Jens; Hartmann, Roland K; Prohaska, Sonja J; Stadler, Peter F

    2014-01-01

    The elucidation of orthology relationships is an important step both in gene function prediction as well as towards understanding patterns of sequence evolution. Orthology assignments are usually derived directly from sequence similarities for large data because more exact approaches exhibit too high computational costs. Here we present PoFF, an extension for the standalone tool Proteinortho, which enhances orthology detection by combining clustering, sequence similarity, and synteny. In the course of this work, FFAdj-MCS, a heuristic that assesses pairwise gene order using adjacencies (a similarity measure related to the breakpoint distance) was adapted to support multiple linear chromosomes and extended to detect duplicated regions. PoFF largely reduces the number of false positives and enables more fine-grained predictions than purely similarity-based approaches. The extension maintains the low memory requirements and the efficient concurrency options of its basis Proteinortho, making the software applicable to very large datasets.

  15. Orthology detection combining clustering and synteny for very large datasets.

    Directory of Open Access Journals (Sweden)

    Marcus Lechner

    Full Text Available The elucidation of orthology relationships is an important step both in gene function prediction as well as towards understanding patterns of sequence evolution. Orthology assignments are usually derived directly from sequence similarities for large data because more exact approaches exhibit too high computational costs. Here we present PoFF, an extension for the standalone tool Proteinortho, which enhances orthology detection by combining clustering, sequence similarity, and synteny. In the course of this work, FFAdj-MCS, a heuristic that assesses pairwise gene order using adjacencies (a similarity measure related to the breakpoint distance was adapted to support multiple linear chromosomes and extended to detect duplicated regions. PoFF largely reduces the number of false positives and enables more fine-grained predictions than purely similarity-based approaches. The extension maintains the low memory requirements and the efficient concurrency options of its basis Proteinortho, making the software applicable to very large datasets.

  16. Comprehensive comparison of large-scale tissue expression datasets

    DEFF Research Database (Denmark)

    Santos Delgado, Alberto; Tsafou, Kalliopi; Stolte, Christian

    2015-01-01

    a comprehensive evaluation of tissue expression data from a variety of experimental techniques and show that these agree surprisingly well with each other and with results from literature curation and text mining. We further found that most datasets support the assumed but not demonstrated distinction between......For tissues to carry out their functions, they rely on the right proteins to be present. Several high-throughput technologies have been used to map out which proteins are expressed in which tissues; however, the data have not previously been systematically compared and integrated. We present......://tissues.jensenlab.org), which makes all the scored and integrated data available through a single user-friendly web interface....

  17. Exploring drivers of wetland hydrologic fluxes across parameters and space

    Science.gov (United States)

    Jones, C. N.; Cheng, F. Y.; Mclaughlin, D. L.; Basu, N. B.; Lang, M.; Alexander, L. C.

    2017-12-01

    Depressional wetlands provide diverse ecosystem services, ranging from critical habitat to the regulation of landscape hydrology. The latter is of particular interest, because while hydrologic connectivity between depressional wetlands and downstream waters has been a focus of both scientific research and policy, it remains difficult to quantify the mode, magnitude, and timing of this connectivity at varying spatial and temporary scales. To do so requires robust empirical and modeling tools that accurately represent surface and subsurface flowpaths between depressional wetlands and other landscape elements. Here, we utilize a parsimonious wetland hydrology model to explore drivers of wetland water fluxes in different archetypal wetland-rich landscapes. We validated the model using instrumented sites from regions that span North America: Prairie Pothole Region (south-central Canada), Delmarva Peninsula (Mid-Atlantic Coastal Plain), and Big Cypress Swamp (southern Florida). Then, using several national scale datasets (e.g., National Wetlands Inventory, USFWS; National Hydrography Dataset, USGS; Soil Survey Geographic Database, NRCS), we conducted a global sensitivity analysis to elucidate dominant drivers of simulated fluxes. Finally, we simulated and compared wetland hydrology in five contrasting landscapes dominated by depressional wetlands: prairie potholes, Carolina and Delmarva bays, pocosins, western vernal pools, and Texas coastal prairie wetlands. Results highlight specific drivers that vary across these regions. Largely, hydroclimatic variables (e.g., PET/P ratios) controlled the timing and magnitude of wetland connectivity, whereas both wetland morphology (e.g., storage capacity and watershed size) and soil characteristics (e.g., ksat and confining layer depth) controlled the duration and mode (surface vs. subsurface) of wetland connectivity. Improved understanding of the drivers of wetland hydrologic connectivity supports enhanced, region

  18. The SAIL databank: linking multiple health and social care datasets

    Directory of Open Access Journals (Sweden)

    Ford David V

    2009-01-01

    Full Text Available Abstract Background Vast amounts of data are collected about patients and service users in the course of health and social care service delivery. Electronic data systems for patient records have the potential to revolutionise service delivery and research. But in order to achieve this, it is essential that the ability to link the data at the individual record level be retained whilst adhering to the principles of information governance. The SAIL (Secure Anonymised Information Linkage databank has been established using disparate datasets, and over 500 million records from multiple health and social care service providers have been loaded to date, with further growth in progress. Methods Having established the infrastructure of the databank, the aim of this work was to develop and implement an accurate matching process to enable the assignment of a unique Anonymous Linking Field (ALF to person-based records to make the databank ready for record-linkage research studies. An SQL-based matching algorithm (MACRAL, Matching Algorithm for Consistent Results in Anonymised Linkage was developed for this purpose. Firstly the suitability of using a valid NHS number as the basis of a unique identifier was assessed using MACRAL. Secondly, MACRAL was applied in turn to match primary care, secondary care and social services datasets to the NHS Administrative Register (NHSAR, to assess the efficacy of this process, and the optimum matching technique. Results The validation of using the NHS number yielded specificity values > 99.8% and sensitivity values > 94.6% using probabilistic record linkage (PRL at the 50% threshold, and error rates were Conclusion With the infrastructure that has been put in place, the reliable matching process that has been developed enables an ALF to be consistently allocated to records in the databank. The SAIL databank represents a research-ready platform for record-linkage studies.

  19. Analysis of Public Datasets for Wearable Fall Detection Systems.

    Science.gov (United States)

    Casilari, Eduardo; Santoyo-Ramón, José-Antonio; Cano-García, José-Manuel

    2017-06-27

    Due to the boom of wireless handheld devices such as smartwatches and smartphones, wearable Fall Detection Systems (FDSs) have become a major focus of attention among the research community during the last years. The effectiveness of a wearable FDS must be contrasted against a wide variety of measurements obtained from inertial sensors during the occurrence of falls and Activities of Daily Living (ADLs). In this regard, the access to public databases constitutes the basis for an open and systematic assessment of fall detection techniques. This paper reviews and appraises twelve existing available data repositories containing measurements of ADLs and emulated falls envisaged for the evaluation of fall detection algorithms in wearable FDSs. The analysis of the found datasets is performed in a comprehensive way, taking into account the multiple factors involved in the definition of the testbeds deployed for the generation of the mobility samples. The study of the traces brings to light the lack of a common experimental benchmarking procedure and, consequently, the large heterogeneity of the datasets from a number of perspectives (length and number of samples, typology of the emulated falls and ADLs, characteristics of the test subjects, features and positions of the sensors, etc.). Concerning this, the statistical analysis of the samples reveals the impact of the sensor range on the reliability of the traces. In addition, the study evidences the importance of the selection of the ADLs and the need of categorizing the ADLs depending on the intensity of the movements in order to evaluate the capability of a certain detection algorithm to discriminate falls from ADLs.

  20. Clusterflock: a flocking algorithm for isolating congruent phylogenomic datasets.

    Science.gov (United States)

    Narechania, Apurva; Baker, Richard; DeSalle, Rob; Mathema, Barun; Kolokotronis, Sergios-Orestis; Kreiswirth, Barry; Planet, Paul J

    2016-10-24

    Collective animal behavior, such as the flocking of birds or the shoaling of fish, has inspired a class of algorithms designed to optimize distance-based clusters in various applications, including document analysis and DNA microarrays. In a flocking model, individual agents respond only to their immediate environment and move according to a few simple rules. After several iterations the agents self-organize, and clusters emerge without the need for partitional seeds. In addition to its unsupervised nature, flocking offers several computational advantages, including the potential to reduce the number of required comparisons. In the tool presented here, Clusterflock, we have implemented a flocking algorithm designed to locate groups (flocks) of orthologous gene families (OGFs) that share an evolutionary history. Pairwise distances that measure phylogenetic incongruence between OGFs guide flock formation. We tested this approach on several simulated datasets by varying the number of underlying topologies, the proportion of missing data, and evolutionary rates, and show that in datasets containing high levels of missing data and rate heterogeneity, Clusterflock outperforms other well-established clustering techniques. We also verified its utility on a known, large-scale recombination event in Staphylococcus aureus. By isolating sets of OGFs with divergent phylogenetic signals, we were able to pinpoint the recombined region without forcing a pre-determined number of groupings or defining a pre-determined incongruence threshold. Clusterflock is an open-source tool that can be used to discover horizontally transferred genes, recombined areas of chromosomes, and the phylogenetic 'core' of a genome. Although we used it here in an evolutionary context, it is generalizable to any clustering problem. Users can write extensions to calculate any distance metric on the unit interval, and can use these distances to 'flock' any type of data.

  1. The SAIL databank: linking multiple health and social care datasets.

    Science.gov (United States)

    Lyons, Ronan A; Jones, Kerina H; John, Gareth; Brooks, Caroline J; Verplancke, Jean-Philippe; Ford, David V; Brown, Ginevra; Leake, Ken

    2009-01-16

    Vast amounts of data are collected about patients and service users in the course of health and social care service delivery. Electronic data systems for patient records have the potential to revolutionise service delivery and research. But in order to achieve this, it is essential that the ability to link the data at the individual record level be retained whilst adhering to the principles of information governance. The SAIL (Secure Anonymised Information Linkage) databank has been established using disparate datasets, and over 500 million records from multiple health and social care service providers have been loaded to date, with further growth in progress. Having established the infrastructure of the databank, the aim of this work was to develop and implement an accurate matching process to enable the assignment of a unique Anonymous Linking Field (ALF) to person-based records to make the databank ready for record-linkage research studies. An SQL-based matching algorithm (MACRAL, Matching Algorithm for Consistent Results in Anonymised Linkage) was developed for this purpose. Firstly the suitability of using a valid NHS number as the basis of a unique identifier was assessed using MACRAL. Secondly, MACRAL was applied in turn to match primary care, secondary care and social services datasets to the NHS Administrative Register (NHSAR), to assess the efficacy of this process, and the optimum matching technique. The validation of using the NHS number yielded specificity values > 99.8% and sensitivity values > 94.6% using probabilistic record linkage (PRL) at the 50% threshold, and error rates were SAIL databank represents a research-ready platform for record-linkage studies.

  2. Analysis of Public Datasets for Wearable Fall Detection Systems

    Directory of Open Access Journals (Sweden)

    Eduardo Casilari

    2017-06-01

    Full Text Available Due to the boom of wireless handheld devices such as smartwatches and smartphones, wearable Fall Detection Systems (FDSs have become a major focus of attention among the research community during the last years. The effectiveness of a wearable FDS must be contrasted against a wide variety of measurements obtained from inertial sensors during the occurrence of falls and Activities of Daily Living (ADLs. In this regard, the access to public databases constitutes the basis for an open and systematic assessment of fall detection techniques. This paper reviews and appraises twelve existing available data repositories containing measurements of ADLs and emulated falls envisaged for the evaluation of fall detection algorithms in wearable FDSs. The analysis of the found datasets is performed in a comprehensive way, taking into account the multiple factors involved in the definition of the testbeds deployed for the generation of the mobility samples. The study of the traces brings to light the lack of a common experimental benchmarking procedure and, consequently, the large heterogeneity of the datasets from a number of perspectives (length and number of samples, typology of the emulated falls and ADLs, characteristics of the test subjects, features and positions of the sensors, etc.. Concerning this, the statistical analysis of the samples reveals the impact of the sensor range on the reliability of the traces. In addition, the study evidences the importance of the selection of the ADLs and the need of categorizing the ADLs depending on the intensity of the movements in order to evaluate the capability of a certain detection algorithm to discriminate falls from ADLs.

  3. Force sensor using changes in magnetic flux

    Science.gov (United States)

    Pickens, Herman L. (Inventor); Richard, James A. (Inventor)

    2012-01-01

    A force sensor includes a magnetostrictive material and a magnetic field generator positioned in proximity thereto. A magnetic field is induced in and surrounding the magnetostrictive material such that lines of magnetic flux pass through the magnetostrictive material. A sensor positioned in the vicinity of the magnetostrictive material measures changes in one of flux angle and flux density when the magnetostrictive material experiences an applied force that is aligned with the lines of magnetic flux.

  4. A high quality finger vascular pattern dataset collected using a custom designed capturing device

    NARCIS (Netherlands)

    Ton, B.T.; Veldhuis, Raymond N.J.

    2013-01-01

    The number of finger vascular pattern datasets available for the research community is scarce, therefore a new finger vascular pattern dataset containing 1440 images is prsented. This dataset is unique in its kind as the images are of high resolution and have a known pixel density. Furthermore this

  5. Something From Nothing (There): Collecting Global IPv6 Datasets from DNS

    NARCIS (Netherlands)

    Fiebig, T.; Borgolte, Kevin; Hao, Shuang; Kruegel, Christopher; Vigna, Giovanny; Spring, Neil; Riley, George F.

    2017-01-01

    Current large-scale IPv6 studies mostly rely on non-public datasets, asmost public datasets are domain specific. For instance, traceroute-based datasetsare biased toward network equipment. In this paper, we present a new methodologyto collect IPv6 address datasets that does not require access to

  6. Reluctance motor employing superconducting magnetic flux switches

    International Nuclear Information System (INIS)

    Spyker, R.L.; Ruckstadter, E.J.

    1992-01-01

    This paper reports that superconducting flux switches controlling the magnetic flux in the poles of a motor will enable the implementation of a reluctance motor using one central single phase winding. A superconducting flux switch consists of a ring of superconducting material surrounding a ferromagnetic pole of the motor. When in the superconducting state the switch will block all magnetic flux attempting to flow in the ferromagnetic core. When switched to the normal state the superconducting switch will allow the magnetic flux to flow freely in that pole. By using one high turns-count coil as a flux generator, and selectively channeling flux among the various poles using the superconducting flux switch, 3-phase operation can be emulated with a single-hase central AC source. The motor will also operate when the flux generating coil is driven by a DC current, provided the magnetic flux switches see a continuously varying magnetic flux. Rotor rotation provides this varying flux due to the change in stator pole inductance it produces

  7. Preliminary results of an attempt to provide soil moisture datasets in order to verify numerical weather prediction models

    International Nuclear Information System (INIS)

    Cassardo, C.; Loglisci, N.

    2005-01-01

    In the recent years, there has been a significant growth in the recognition of the soil moisture importance in large-scale hydrology and climate modelling. Soil moisture is a lower boundary condition, which rules the partitioning of energy in terms of sensible and latent heat flux. Wrong estimations of soil moisture lead to wrong simulation of the surface layer evolution and hence precipitations and cloud cover forecasts could be consequently affected. This is true for large scale medium-range weather forecasts as well as for local-scale short range weather forecasts, particularly in those situations in which local convection is well developed. Unfortunately; despite the importance of this physical parameter there are only few soil moisture data sets sparse in time and in space around in the world. Due to this scarcity of soil moisture observations, we developed an alternative method to provide soil moisture datasets in order to verify numerical weather prediction models. In this paper are presented the preliminary results of an attempt to verify soil moisture fields predicted by a mesoscale model. The data for the comparison were provided by the simulations of the diagnostic land surface scheme LSPM (Land Surface Process Model), widely used at the Piedmont Regional Weather Service for agro-meteorological purposes. To this end, LSPM was initialized and driven by Synop observations, while the surface (vegetation and soil) parameter values were initialized by ECOCLIMAP global dataset at 1km 2 resolution

  8. Inverse modelling estimates of N2O surface emissions and stratospheric losses using a global dataset

    Science.gov (United States)

    Thompson, R. L.; Bousquet, P.; Chevallier, F.; Dlugokencky, E. J.; Vermeulen, A. T.; Aalto, T.; Haszpra, L.; Meinhardt, F.; O'Doherty, S.; Moncrieff, J. B.; Popa, M.; Steinbacher, M.; Jordan, A.; Schuck, T. J.; Brenninkmeijer, C. A.; Wofsy, S. C.; Kort, E. A.

    2010-12-01

    Nitrous oxide (N2O) levels have been steadily increasing in the atmosphere over the past few decades at a rate of approximately 0.3% per year. This trend is of major concern as N2O is both a long-lived Greenhouse Gas (GHG) and an Ozone Depleting Substance (ODS), as it is a precursor of NO and NO2, which catalytically destroy ozone in the stratosphere. Recently, N2O emissions have been recognised as the most important ODS emissions and are now of greater importance than emissions of CFC's. The growth in atmospheric N2O is predominantly due to the enhancement of surface emissions by human activities. Most notably, the intensification and proliferation of agriculture since the mid-19th century, which has been accompanied by the increased input of reactive nitrogen to soils and has resulted in significant perturbations to the natural N-cycle and emissions of N2O. There exist two approaches for estimating N2O emissions, the so-called 'bottom-up' and 'top-down' approaches. Top-down approaches, based on the inversion of atmospheric measurements, require an estimate of the loss of N2O via photolysis and oxidation in the stratosphere. Uncertainties in the loss magnitude contribute uncertainties of 15 to 20% to the global annual surface emissions, complicating direct comparisons between bottom-up and top-down estimates. In this study, we present a novel inversion framework for the simultaneous optimization of N2O surface emissions and the magnitude of the loss, which avoids errors in the emissions due to incorrect assumptions about the lifetime of N2O. We use a Bayesian inversion with a variational formulation (based on 4D-Var) in order to handle very large datasets. N2O fluxes are retrieved at 4-weekly resolution over a global domain with a spatial resolution of 3.75° x 2.5° longitude by latitude. The efficacy of the simultaneous optimization of emissions and losses is tested using a global synthetic dataset, which mimics the available atmospheric data. Lastly, using real

  9. Flux of Cadmium through Euphausiids

    International Nuclear Information System (INIS)

    Benayoun, G.; Fowler, S.W.; Oregioni, B.

    1976-01-01

    Flux of the heavy metal cadmium through the euphausiid Meganyctiphanes norvegica was examined. Radiotracer experiments showed that cadmium can be accumulated either directly from water or through the food chain. When comparing equilibrium cadmium concentration factors based on stable element measurements with those obtained from radiotracer experiments, it is evident that exchange between cadmium in the water and that in euphausiid tissue is a relatively slow process, indicating that, in the long term, ingestion of cadmium will probably be the more important route for the accumulation of this metal. Approximately 10% of cadmium ingested by euphausiids was incorporated into internal tissues when the food source was radioactive Artemia. After 1 month cadmium, accumulated directly from water, was found to be most concentrated in the viscera with lesser amounts in eyes, exoskeleton and muscle, respectively. Use of a simple model, based on the assumption that cadmium taken in by the organism must equal cadmium released plus that accumulated in tissue, allowed assessment of the relative importance of various metabolic parameters in controlling the cadmium flux through euphausiids. Fecal pellets, due to their relatively high rate of production and high cadmium content, accounted for 84% of the total cadmium flux through M. norvegica. Comparisons of stable cadmium concentrations in natural euphausiid food and the organism's resultant fecal pellets indicate that the cadmium concentration in ingested material was increased nearly 5-fold during its passage through the euphausiid. From comparisons of all routes by which cadmium can be released from M. norvegica to the water column, it is concluded that fecal pellet deposition represents the principal mechanism effecting the downward vertical transport of cadmium by this species. (author)

  10. Framework for Flux Qubit Design

    Science.gov (United States)

    Yan, Fei; Kamal, Archana; Krantz, Philip; Campbell, Daniel; Kim, David; Yoder, Jonilyn; Orlando, Terry; Gustavsson, Simon; Oliver, William; Engineering Quantum Systems Team

    A qubit design for higher performance relies on the understanding of how various qubit properties are related to design parameters. We construct a framework for understanding the qubit design in the flux regime. We explore different parameter regimes, looking for features desirable for certain purpose in the context of quantum computing. This research was funded by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA) via MIT Lincoln Laboratory under Air Force Contract No. FA8721-05-C-0002.

  11. FSFE: Fake Spectra Flux Extractor

    Science.gov (United States)

    Bird, Simeon

    2017-10-01

    The fake spectra flux extractor generates simulated quasar absorption spectra from a particle or adaptive mesh-based hydrodynamic simulation. It is implemented as a python module. It can produce both hydrogen and metal line spectra, if the simulation includes metals. The cloudy table for metal ionization fractions is included. Unlike earlier spectral generation codes, it produces absorption from each particle close to the sight-line individually, rather than first producing an average density in each spectral pixel, thus substantially preserving more of the small-scale velocity structure of the gas. The code supports both Gadget (ascl:0003.001) and AREPO.

  12. Four-collector flux sensor

    International Nuclear Information System (INIS)

    Wiegand, W.J. Jr.; Bullis, R.H.; Mongeon, R.J.

    1980-01-01

    A flowmeter based on ion drift techniques was developed for measuring the rate of flow of a fluid through a given cross-section. Ion collectors are positioned on each side of an immediately adjacent to ion source. When air flows axially through the region in which ions are produced and appropriate electric fields are maintained between the collectors, an electric current flows to each collector due to the net motion of the ions. The electric currents and voltages and other parameters which define the flow are combined in an electric circuit so that the flux of the fluid can be determined. (DN)

  13. Privacy preserving data anonymization of spontaneous ADE reporting system dataset.

    Science.gov (United States)

    Lin, Wen-Yang; Yang, Duen-Chuan; Wang, Jie-Teng

    2016-07-18

    To facilitate long-term safety surveillance of marketing drugs, many spontaneously reporting systems (SRSs) of ADR events have been established world-wide. Since the data collected by SRSs contain sensitive personal health information that should be protected to prevent the identification of individuals, it procures the issue of privacy preserving data publishing (PPDP), that is, how to sanitize (anonymize) raw data before publishing. Although much work has been done on PPDP, very few studies have focused on protecting privacy of SRS data and none of the anonymization methods is favorable for SRS datasets, due to which contain some characteristics such as rare events, multiple individual records, and multi-valued sensitive attributes. We propose a new privacy model called MS(k, θ (*) )-bounding for protecting published spontaneous ADE reporting data from privacy attacks. Our model has the flexibility of varying privacy thresholds, i.e., θ (*) , for different sensitive values and takes the characteristics of SRS data into consideration. We also propose an anonymization algorithm for sanitizing the raw data to meet the requirements specified through the proposed model. Our algorithm adopts a greedy-based clustering strategy to group the records into clusters, conforming to an innovative anonymization metric aiming to minimize the privacy risk as well as maintain the data utility for ADR detection. Empirical study was conducted using FAERS dataset from 2004Q1 to 2011Q4. We compared our model with four prevailing methods, including k-anonymity, (X, Y)-anonymity, Multi-sensitive l-diversity, and (α, k)-anonymity, evaluated via two measures, Danger Ratio (DR) and Information Loss (IL), and considered three different scenarios of threshold setting for θ (*) , including uniform setting, level-wise setting and frequency-based setting. We also conducted experiments to inspect the impact of anonymized data on the strengths of discovered ADR signals. With all three

  14. Two MODIS Aerosol Products over Ocean on the Terra and Aqua CERES SSF Datasets.

    Science.gov (United States)

    Ignatov, Alexander; Minnis, Patrick; Loeb, Norman; Wielicki, Bruce; Miller, Walter; Sun-Mack, Sunny; Tanré, Didier; Remer, Lorraine; Laszlo, Istvan; Geier, Erika

    2005-04-01

    Understanding the impact of aerosols on the earth's radiation budget and the long-term climate record requires consistent measurements of aerosol properties and radiative fluxes. The Clouds and the Earth's Radiant Energy System (CERES) Science Team combines satellite-based retrievals of aerosols, clouds, and radiative fluxes into Single Scanner Footprint (SSF) datasets from the Terra and Aqua satellites. Over ocean, two aerosol products are derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) using different sampling and aerosol algorithms. The primary, or M, product is taken from the standard multispectral aerosol product developed by the MODIS aerosol group while a simpler, secondary [Advanced Very High Resolution Radiometer (AVHRR) like], or A, product is derived by the CERES Science Team using a different cloud clearing method and a single-channel aerosol algorithm. Two aerosol optical depths (AOD), τA1 and τA2, are derived from MODIS bands 1 (0.644 μm) and 6 (1.632 μm) resembling the AVHRR/3 channels 1 and 3A, respectively. On Aqua the retrievals are made in band 7 (2.119 μm) because of poor quality data from band 6. The respective Ångström exponents can be derived from the values of τ. The A product serves as a backup for the M product. More importantly, the overlap of these aerosol products is essential for placing the 20+ year heritage AVHRR aerosol record in the context of more advanced aerosol sensors and algorithms such as that used for the M product.This study documents the M and A products, highlighting their CERES SSF specifics. Based on 2 weeks of global Terra data, coincident M and A AODs are found to be strongly correlated in both bands. However, both domains in which the M and A aerosols are available, and the respective τ/α statistics significantly differ because of discrepancies in sampling due to differences in cloud and sun-glint screening. In both aerosol products, correlation is observed between the retrieved

  15. Global Bedload Flux Modeling and Analysis in Large Rivers

    Science.gov (United States)

    Islam, M. T.; Cohen, S.; Syvitski, J. P.

    2017-12-01

    Proper sediment transport quantification has long been an area of interest for both scientists and engineers in the fields of geomorphology, and management of rivers and coastal waters. Bedload flux is important for monitoring water quality and for sustainable development of coastal and marine bioservices. Bedload measurements, especially for large rivers, is extremely scarce across time, and many rivers have never been monitored. Bedload measurements in rivers, is particularly acute in developing countries where changes in sediment yields is high. The paucity of bedload measurements is the result of 1) the nature of the problem (large spatial and temporal uncertainties), and 2) field costs including the time-consuming nature of the measurement procedures (repeated bedform migration tracking, bedload samplers). Here we present a first of its kind methodology for calculating bedload in large global rivers (basins are >1,000 km. Evaluation of model skill is based on 113 bedload measurements. The model predictions are compared with an empirical model developed from the observational dataset in an attempt to evaluate the differences between a physically-based numerical model and a lumped relationship between bedload flux and fluvial and basin parameters (e.g., discharge, drainage area, lithology). The initial study success opens up various applications to global fluvial geomorphology (e.g. including the relationship between suspended sediment (wash load) and bedload). Simulated results with known uncertainties offers a new research product as a valuable resource for the whole scientific community.

  16. Airflows and turbulent flux measurements in mountainous terrain: Part 1. Canopy and local effects

    Science.gov (United States)

    Turnipseed, Andrew A.; Anderson, Dean E.; Blanken, Peter D.; Baugh, William M.; Monson, Russell K.

    2003-01-01

    We have studied the effects of local topography and canopy structure on turbulent flux measurements at a site located in mountainous terrain within a subalpine, coniferous forest. Our primary aim was to determine whether the complex terrain of the site affects the accuracy of eddy flux measurements from a practical perspective. We observed displacement heights, roughness lengths, spectral peaks, turbulent length scales, and profiles of turbulent intensities that were comparable in magnitude and pattern to those reported for forest canopies in simpler terrain. We conclude that in many of these statistical measures, the local canopy exerts considerably more influence than does topographical complexity. Lack of vertical flux divergence and modeling suggests that the flux footprints for the site are within the standards acceptable for the application of flux statistics. We investigated three different methods of coordinate rotation: double rotation (DR), triple rotation (TR), and planar-fit rotation (PF). Significant variability in rotation angles at low wind speeds was encountered with the commonly used DR and TR methods, as opposed to the PF method, causing some overestimation of the fluxes. However, these differences in fluxes were small when applied to large datasets involving sensible heat and CO2 fluxes. We observed evidence of frequent drainage flows near the ground during stable, stratified conditions at night. Concurrent with the appearance of these flows, we observed a positive bias in the mean vertical wind speed, presumably due to subtle topographic variations inducing a flow convergence below the measurement sensors. In the presence of such drainage flows, advection of scalars and non-zero bias in the mean vertical wind speed can complicate closure of the mass conservation budget at the site.

  17. Standardization of GIS datasets for emergency preparedness of NPPs

    International Nuclear Information System (INIS)

    Saindane, Shashank S.; Suri, M.M.K.; Otari, Anil; Pradeepkumar, K.S.

    2012-01-01

    Probability of a major nuclear accident which can lead to large scale release of radioactivity into environment is extremely small by the incorporation of safety systems and defence-in-depth philosophy. Nevertheless emergency preparedness for implementation of counter measures to reduce the consequences are required for all major nuclear facilities. Iodine prophylaxis, Sheltering, evacuation etc. are protective measures to be implemented for members of public in the unlikely event of any significant releases from nuclear facilities. Bhabha Atomic Research Centre has developed a GIS supported Nuclear Emergency Preparedness Program. Preparedness for Response to Nuclear emergencies needs geographical details of the affected locations specially Nuclear Power Plant Sites and nearby public domain. Geographical information system data sets which the planners are looking for will have appropriate details in order to take decision and mobilize the resources in time and follow the Standard Operating Procedures. Maps are 2-dimensional representations of our real world and GIS makes it possible to manipulate large amounts of geo-spatially referenced data and convert it into information. This has become an integral part of the nuclear emergency preparedness and response planning. This GIS datasets consisting of layers such as village settlements, roads, hospitals, police stations, shelters etc. is standardized and effectively used during the emergency. The paper focuses on the need of standardization of GIS datasets which in turn can be used as a tool to display and evaluate the impact of standoff distances and selected zones in community planning. It will also highlight the database specifications which will help in fast processing of data and analysis to derive useful and helpful information. GIS has the capability to store, manipulate, analyze and display the large amount of required spatial and tabular data. This study intends to carry out a proper response and preparedness

  18. Forest restoration: a global dataset for biodiversity and vegetation structure.

    Science.gov (United States)

    Crouzeilles, Renato; Ferreira, Mariana S; Curran, Michael

    2016-08-01

    Restoration initiatives are becoming increasingly applied around the world. Billions of dollars have been spent on ecological restoration research and initiatives, but restoration outcomes differ widely among these initiatives in part due to variable socioeconomic and ecological contexts. Here, we present the most comprehensive dataset gathered to date on forest restoration. It encompasses 269 primary studies across 221 study landscapes in 53 countries and contains 4,645 quantitative comparisons between reference ecosystems (e.g., old-growth forest) and degraded or restored ecosystems for five taxonomic groups (mammals, birds, invertebrates, herpetofauna, and plants) and five measures of vegetation structure reflecting different ecological processes (cover, density, height, biomass, and litter). We selected studies that (1) were conducted in forest ecosystems; (2) had multiple replicate sampling sites to measure indicators of biodiversity and/or vegetation structure in reference and restored and/or degraded ecosystems; and (3) used less-disturbed forests as a reference to the ecosystem under study. We recorded (1) latitude and longitude; (2) study year; (3) country; (4) biogeographic realm; (5) past disturbance type; (6) current disturbance type; (7) forest conversion class; (8) restoration activity; (9) time that a system has been disturbed; (10) time elapsed since restoration started; (11) ecological metric used to assess biodiversity; and (12) quantitative value of the ecological metric of biodiversity and/or vegetation structure for reference and restored and/or degraded ecosystems. These were the most common data available in the selected studies. We also estimated forest cover and configuration in each study landscape using a recently developed 1 km consensus land cover dataset. We measured forest configuration as the (1) mean size of all forest patches; (2) size of the largest forest patch; and (3) edge:area ratio of forest patches. Global analyses of the

  19. BanglaLekha-Isolated: A multi-purpose comprehensive dataset of Handwritten Bangla Isolated characters

    Directory of Open Access Journals (Sweden)

    Mithun Biswas

    2017-06-01

    Full Text Available BanglaLekha-Isolated, a Bangla handwritten isolated character dataset is presented in this article. This dataset contains 84 different characters comprising of 50 Bangla basic characters, 10 Bangla numerals and 24 selected compound characters. 2000 handwriting samples for each of the 84 characters were collected, digitized and pre-processed. After discarding mistakes and scribbles, 1,66,105 handwritten character images were included in the final dataset. The dataset also includes labels indicating the age and the gender of the subjects from whom the samples were collected. This dataset could be used not only for optical handwriting recognition research but also to explore the influence of gender and age on handwriting. The dataset is publicly available at https://data.mendeley.com/datasets/hf6sf8zrkc/2.

  20. A Research Graph dataset for connecting research data repositories using RD-Switchboard.

    Science.gov (United States)

    Aryani, Amir; Poblet, Marta; Unsworth, Kathryn; Wang, Jingbo; Evans, Ben; Devaraju, Anusuriya; Hausstein, Brigitte; Klas, Claus-Peter; Zapilko, Benjamin; Kaplun, Samuele

    2018-05-29

    This paper describes the open access graph dataset that shows the connections between Dryad, CERN, ANDS and other international data repositories to publications and grants across multiple research data infrastructures. The graph dataset was created using the Research Graph data model and the Research Data Switchboard (RD-Switchboard), a collaborative project by the Research Data Alliance DDRI Working Group (DDRI WG) with the aim to discover and connect the related research datasets based on publication co-authorship or jointly funded grants. The graph dataset allows researchers to trace and follow the paths to understanding a body of work. By mapping the links between research datasets and related resources, the graph dataset improves both their discovery and visibility, while avoiding duplicate efforts in data creation. Ultimately, the linked datasets may spur novel ideas, facilitate reproducibility and re-use in new applications, stimulate combinatorial creativity, and foster collaborations across institutions.

  1. Testing the applicability of BIOME-BGC to simulate beech gross primary production in Europe using a new continental weather dataset

    DEFF Research Database (Denmark)

    Chiesi, Marta; Chirici, Gherardo; Marchetti, Marco

    2016-01-01

    A daily 1-km Pan-European weather dataset can drive the BIOME-BGC model for the estimation of current and future beech gross primary production (GPP). Annual beech GPP is affected primarily by spring temperature and more irregularly by summer water stress.The spread of beech forests in Europe...... forest ecosystems having different climatic conditions where the eddy covariance technique is used to measure water and carbon fluxes. The experiment is in three main steps. First, the accuracy of BIOME-BGC GPP simulations is assessed through comparison with flux observations. Second, the influence...... of two major meteorological drivers (spring minimum temperature and growing season dryness) on observed and simulated inter-annual GPP variations is analysed. Lastly, the impacts of two climate change scenarios on beech GPP are evaluated through statistical analyses of the ground data and model...

  2. Retrieving SW fluxes from geostationary narrowband radiances for the NASA-CERES SYN1deg product

    Science.gov (United States)

    Wrenn, F. J., IV; Doelling, D. R.; Liang, L.

    2017-12-01

    The CERES mission was designed to measure the natural variability of the net TOA flux over long time scales relevant to climate monitoring. To achieve this goal, CERES provides the level-3 SSF1deg, SYN1deg, and EBAF monthly 1° by 1° regional TOA flux. The single satellite (Terra or Aqua) SSF1deg 24-hour shortwave flux is based on one daytime measurements and assumes constant meteorology to model the diurnal change in albedo. To accurately describe regions with a prominent diurnal signal, the SYN1deg Edition4 dataset employs hourly geostationary (GEO) measurements. This improves upon Edition3, which used 3-hourly GEO measurements and with temporal interpolation. The EBAF product combines the temporal stability of the SSF1deg product with the diurnal information from SYN1deg and removes the CERES instrument calibration bias by constraining the net flux balance to the ocean heat storage term. The SYN-1deg product retrieves hourly SW fluxes from GEO measurements. Over regions with large diurnal cycles, such as maritime stratus and land afternoon convective locations, the GEO derived SW fluxes will capture the diurnal flux not observed with Terra or Aqua sun-synchronous satellites. Obtaining fluxes from geostationary satellite radiance is a multistep process. First, most GEO visible imagers lack calibration and must be calibrated to MODIS and VIIRS. Second, the GEO imager visible channel radiances are converted to broadband radiances using empirical and theoretical models. The lack of coincident, collocated, and co-angled GEO and CERES measurements makes building an empirical model difficult. The narrowband to broadband models are a function of surface and cloud conditions, which are difficult to identify due to the inconsistent cloud retrievals between the 16 GEO imagers used in the CERES record. Third, the GEO derived broadband radiances are passed through the CERES angular distribution model (ADM) to convert the radiances to fluxes. Lastly, the GEO derived

  3. Triode for magnetic flux quanta.

    Science.gov (United States)

    Vlasko-Vlasov, Vitalii; Colauto, Fabiano; Benseman, Timothy; Rosenmann, Daniel; Kwok, Wai-Kwong

    We designed a magnetic vortex triode using an array of closely spaced soft magnetic Py strips on top of a Nb superconducting film. The strips act similar to the grid electrode in an electronic triode, where the electron flow is regulated by the grid potential. In our case, we tune the vortex motion by the magnetic charge potential of the strip edges, using a small magnetic field rotating in the film plane. The magnetic charges emerging at the stripe edges and proportional to the magnetization component perpendicular to the edge direction, form linear potential barriers or valleys for vortex motion in the superconducting layer. We directly imaged the normal flux penetration into the Py/Nb films and observed retarded or accelerated entry of the normal vortices depending on the in-plane magnetization direction in the stripes. The observed flux behavior is explained by interactions between magnetically charged lines and magnetic monopoles of vortices similar to those between electrically charged strings and point charges. We discuss the possibility of using our design for manipulation of individual vortices in high-speed, low-power superconducting electronic circuits. This work was supported by the U.S. DOE, Office of Science, Materials Sciences and Engineering Division, and Office of BES (contract DE-AC02-06CH11357). F. Colauto thanks the Sao Paulo Research Foundation FAPESP (Grant No. 2015/06.085-3).

  4. Neutron flux enhancement at LASREF

    International Nuclear Information System (INIS)

    Sommer, W.F.; Ferguson, P.D.; Wechsler, M.S.

    1991-01-01

    The accelerator at the Los Alamos Meson Physics Facility produces a 1-mA beam of protons at an energy of 800 MeV. Since 1985, the Los Alamos Spallation Radiation Effects Facility (LASREF) has made use of the neutron flux that is generated as the incident protons interact with the nuclei in targets and a copper beam stop. A variety of basic and applied experiments in radiation damage and radiation effects have been completed. Recent studies indicate that the flux at LASREF can be increased by at least a factor of ten from the present level of about 5 E+17 m -2 s -1 . This requires changing the beam-stop material from Cu to W and optimizing the geometry of the beam-target interaction region. These studies are motivated by the need for a large volume, high energy, and high intensity neutron source in the development of materials for advanced energy concepts such as fusion reactors. 18 refs., 7 figs., 2 tabs

  5. Neutron flux enhancement at LASREF

    Energy Technology Data Exchange (ETDEWEB)

    Sommer, W.F. (Los Alamos National Lab., Los Alamos, NM (United States)); Ferguson, P.D. (Univ. of Missouri, Rolla, MO (United States)); Wechsler, M.S. (Iowa State Univ., Ames, IA (United States))

    1992-09-01

    The accelerator at the Los Alamos Meson Physiscs Facility produces a 1 mA beam of protons at an energy of 800 MeV. Since 1985, the Los Alamos Spallation Radiation Effects Facility (LASREF) has made use of the neutron flux that is generated as the incident protons interact with the targets and a copper beam stop. A variety of basic and applied experiments in radiation damage and radiation effects have been completed. Recent studies indicate that the flux at LASREF can be increased by at least a factor of 10 from the present level of about 5 E + 17 m[sup -2] s[sup -1]. This requires changing the beam stop material from Cu to W and optimizing the geometry of the beam-target interaction region. These studies are motivated by the need for a large volume, high energy, and high intensity neutron source in the development of materials for advanced energy concepts such as fusion reactors. (orig.).

  6. Emerging ecological datasets with application for modeling North American dust emissions

    Science.gov (United States)

    McCord, S.; Stauffer, N. G.; Garman, S.; Webb, N.

    2017-12-01

    In 2011 the US Bureau of Land Management (BLM) established the Assessment, Inventory and Monitoring (AIM) program to monitor the condition of BLM land and to provide data to support evidence-based management of multi-use public lands. The monitoring program shares core data collection methods with the Natural Resources Conservation Service's (NRCS) National Resources Inventory (NRI), implemented on private lands nationally. Combined, the two programs have sampled >30,000 locations since 2003 to provide vegetation composition, vegetation canopy height, the size distribution of inter-canopy gaps, soil texture and crusting information on rangelands and pasture lands across North America. The BLM implements AIM on more than 247.3 million acres of land across the western US, encompassing major dust source regions of the Chihuahuan, Sonoran, Mojave and Great Basin deserts, the Colorado Plateau, and potential high-latitude dust sources in Alaska. The AIM data are publicly available and can be used to support modeling of land surface and boundary-layer processes, including dust emission. While understanding US dust source regions and emission processes has been of national interest since the 1930s Dust Bowl, most attention has been directed to the croplands of the Great Plains and emission hot spots like Owens Lake, California. The magnitude, spatial extent and temporal dynamics of dust emissions from western dust source areas remain highly uncertain. Here, we use ensemble modeling with empirical and physically-based dust emission schemes applied to AIM monitoring data to assess regional-scale patterns of aeolian sediment mass fluxes and dust emissions. The analysis enables connections to be made between dust emission rates at source and other indicators of ecosystem function at the landscape scale. Emerging ecological datasets like AIM provide new opportunities to evaluate aeolian sediment transport responses to land surface conditions, potential interactions with

  7. Provenance of Earth Science Datasets - How Deep Should One Go?

    Science.gov (United States)

    Ramapriyan, H.; Manipon, G. J. M.; Aulenbach, S.; Duggan, B.; Goldstein, J.; Hua, H.; Tan, D.; Tilmes, C.; Wilson, B. D.; Wolfe, R.; Zednik, S.

    2015-12-01

    For credibility of scientific research, transparency and reproducibility are essential. This fundamental tenet has been emphasized for centuries, and has been receiving increased attention in recent years. The Office of Management and Budget (2002) addressed reproducibility and other aspects of quality and utility of information from federal agencies. Specific guidelines from NASA (2002) are derived from the above. According to these guidelines, "NASA requires a higher standard of quality for information that is considered influential. Influential scientific, financial, or statistical information is defined as NASA information that, when disseminated, will have or does have clear and substantial impact on important public policies or important private sector decisions." For information to be compliant, "the information must be transparent and reproducible to the greatest possible extent." We present how the principles of transparency and reproducibility have been applied to NASA data supporting the Third National Climate Assessment (NCA3). The depth of trace needed of provenance of data used to derive conclusions in NCA3 depends on how the data were used (e.g., qualitatively or quantitatively). Given that the information is diligently maintained in the agency archives, it is possible to trace from a figure in the publication through the datasets, specific files, algorithm versions, instruments used for data collection, and satellites, as well as the individuals and organizations involved in each step. Such trace back permits transparency and reproducibility.

  8. A dataset from bottom trawl survey around Taiwan

    Directory of Open Access Journals (Sweden)

    Kwang-tsao Shao

    2012-05-01

    Full Text Available Bottom trawl fishery is one of the most important coastal fisheries in Taiwan both in production and economic values. However, its annual production started to decline due to overfishing since the 1980s. Its bycatch problem also damages the fishery resource seriously. Thus, the government banned the bottom fishery within 3 nautical miles along the shoreline in 1989. To evaluate the effectiveness of this policy, a four year survey was conducted from 2000–2003, in the waters around Taiwan and Penghu (Pescadore Islands, one region each year respectively. All fish specimens collected from trawling were brought back to lab for identification, individual number count and body weight measurement. These raw data have been integrated and established in Taiwan Fish Database (http://fishdb.sinica.edu.tw. They have also been published through TaiBIF (http://taibif.tw, FishBase and GBIF (website see below. This dataset contains 631 fish species and 3,529 records, making it the most complete demersal fish fauna and their temporal and spatial distributional data on the soft marine habitat in Taiwan.

  9. Integrated interpretation of overlapping AEM datasets achieved through standardisation

    Science.gov (United States)

    Sørensen, Camilla C.; Munday, Tim; Heinson, Graham

    2015-12-01

    Numerous airborne electromagnetic surveys have been acquired in Australia using a variety of systems. It is not uncommon to find two or more surveys covering the same ground, but acquired using different systems and at different times. Being able to combine overlapping datasets and get a spatially coherent resistivity-depth image of the ground can assist geological interpretation, particularly when more subtle geophysical responses are important. Combining resistivity-depth models obtained from the inversion of airborne electromagnetic (AEM) data can be challenging, given differences in system configuration, geometry, flying height and preservation or monitoring of system acquisition parameters such as waveform. In this study, we define and apply an approach to overlapping AEM surveys, acquired by fixed wing and helicopter time domain electromagnetic (EM) systems flown in the vicinity of the Goulds Dam uranium deposit in the Frome Embayment, South Australia, with the aim of mapping the basement geometry and the extent of the Billeroo palaeovalley. Ground EM soundings were used to standardise the AEM data, although results indicated that only data from the REPTEM system needed to be corrected to bring the two surveys into agreement and to achieve coherent spatial resistivity-depth intervals.

  10. A global dataset of sub-daily rainfall indices

    Science.gov (United States)

    Fowler, H. J.; Lewis, E.; Blenkinsop, S.; Guerreiro, S.; Li, X.; Barbero, R.; Chan, S.; Lenderink, G.; Westra, S.

    2017-12-01

    It is still uncertain how hydrological extremes will change with global warming as we do not fully understand the processes that cause extreme precipitation under current climate variability. The INTENSE project is using a novel and fully-integrated data-modelling approach to provide a step-change in our understanding of the nature and drivers of global precipitation extremes and change on societally relevant timescales, leading to improved high-resolution climate model representation of extreme rainfall processes. The INTENSE project is in conjunction with the World Climate Research Programme (WCRP)'s Grand Challenge on 'Understanding and Predicting Weather and Climate Extremes' and the Global Water and Energy Exchanges Project (GEWEX) Science questions. A new global sub-daily precipitation dataset has been constructed (data collection is ongoing). Metadata for each station has been calculated, detailing record lengths, missing data, station locations. A set of global hydroclimatic indices have been produced based upon stakeholder recommendations including indices that describe maximum rainfall totals and timing, the intensity, duration and frequency of storms, frequency of storms above specific thresholds and information about the diurnal cycle. This will provide a unique global data resource on sub-daily precipitation whose derived indices will be freely available to the wider scientific community.

  11. The Centennial Trends Greater Horn of Africa precipitation dataset

    Science.gov (United States)

    Funk, Chris; Nicholson, Sharon E.; Landsfeld, Martin F.; Klotter, Douglas; Peterson, Pete J.; Harrison, Laura

    2015-01-01

    East Africa is a drought prone, food and water insecure region with a highly variable climate. This complexity makes rainfall estimation challenging, and this challenge is compounded by low rain gauge densities and inhomogeneous monitoring networks. The dearth of observations is particularly problematic over the past decade, since the number of records in globally accessible archives has fallen precipitously. This lack of data coincides with an increasing scientific and humanitarian need to place recent seasonal and multi-annual East African precipitation extremes in a deep historic context. To serve this need, scientists from the UC Santa Barbara Climate Hazards Group and Florida State University have pooled their station archives and expertise to produce a high quality gridded ‘Centennial Trends’ precipitation dataset. Additional observations have been acquired from the national meteorological agencies and augmented with data provided by other universities. Extensive quality control of the data was carried out and seasonal anomalies interpolated using kriging. This paper documents the CenTrends methodology and data.

  12. Dataset on daytime outdoor thermal comfort for Belo Horizonte, Brazil.

    Science.gov (United States)

    Hirashima, Simone Queiroz da Silveira; Assis, Eleonora Sad de; Nikolopoulou, Marialena

    2016-12-01

    This dataset describe microclimatic parameters of two urban open public spaces in the city of Belo Horizonte, Brazil; physiological equivalent temperature (PET) index values and the related subjective responses of interviewees regarding thermal sensation perception and preference and thermal comfort evaluation. Individuals and behavioral characteristics of respondents were also presented. Data were collected at daytime, in summer and winter, 2013. Statistical treatment of this data was firstly presented in a PhD Thesis ("Percepção sonora e térmica e avaliação de conforto em espaços urbanos abertos do município de Belo Horizonte - MG, Brasil" (Hirashima, 2014) [1]), providing relevant information on thermal conditions in these locations and on thermal comfort assessment. Up to now, this data was also explored in the article "Daytime Thermal Comfort in Urban Spaces: A Field Study in Brazil" (Hirashima et al., in press) [2]. These references are recommended for further interpretation and discussion.

  13. Statistically and Computationally Efficient Estimating Equations for Large Spatial Datasets

    KAUST Repository

    Sun, Ying

    2014-11-07

    For Gaussian process models, likelihood based methods are often difficult to use with large irregularly spaced spatial datasets, because exact calculations of the likelihood for n observations require O(n3) operations and O(n2) memory. Various approximation methods have been developed to address the computational difficulties. In this paper, we propose new unbiased estimating equations based on score equation approximations that are both computationally and statistically efficient. We replace the inverse covariance matrix that appears in the score equations by a sparse matrix to approximate the quadratic forms, then set the resulting quadratic forms equal to their expected values to obtain unbiased estimating equations. The sparse matrix is constructed by a sparse inverse Cholesky approach to approximate the inverse covariance matrix. The statistical efficiency of the resulting unbiased estimating equations are evaluated both in theory and by numerical studies. Our methods are applied to nearly 90,000 satellite-based measurements of water vapor levels over a region in the Southeast Pacific Ocean.

  14. Statistically and Computationally Efficient Estimating Equations for Large Spatial Datasets

    KAUST Repository

    Sun, Ying; Stein, Michael L.

    2014-01-01

    For Gaussian process models, likelihood based methods are often difficult to use with large irregularly spaced spatial datasets, because exact calculations of the likelihood for n observations require O(n3) operations and O(n2) memory. Various approximation methods have been developed to address the computational difficulties. In this paper, we propose new unbiased estimating equations based on score equation approximations that are both computationally and statistically efficient. We replace the inverse covariance matrix that appears in the score equations by a sparse matrix to approximate the quadratic forms, then set the resulting quadratic forms equal to their expected values to obtain unbiased estimating equations. The sparse matrix is constructed by a sparse inverse Cholesky approach to approximate the inverse covariance matrix. The statistical efficiency of the resulting unbiased estimating equations are evaluated both in theory and by numerical studies. Our methods are applied to nearly 90,000 satellite-based measurements of water vapor levels over a region in the Southeast Pacific Ocean.

  15. Challenges and Experiences of Building Multidisciplinary Datasets across Cultures

    Science.gov (United States)

    Jamiyansharav, K.; Laituri, M.; Fernandez-Gimenez, M.; Fassnacht, S. R.; Venable, N. B. H.; Allegretti, A. M.; Reid, R.; Baival, B.; Jamsranjav, C.; Ulambayar, T.; Linn, S.; Angerer, J.

    2017-12-01

    Efficient data sharing and management are key challenges to multidisciplinary scientific research. These challenges are further complicated by adding a multicultural component. We address the construction of a complex database for social-ecological analysis in Mongolia. Funded by the National Science Foundation (NSF) Dynamics of Coupled Natural and Human (CNH) Systems, the Mongolian Rangelands and Resilience (MOR2) project focuses on the vulnerability of Mongolian pastoral systems to climate change and adaptive capacity. The MOR2 study spans over three years of fieldwork in 36 paired districts (Soum) from 18 provinces (Aimag) of Mongolia that covers steppe, mountain forest steppe, desert steppe and eastern steppe ecological zones. Our project team is composed of hydrologists, social scientists, geographers, and ecologists. The MOR2 database includes multiple ecological, social, meteorological, geospatial and hydrological datasets, as well as archives of original data and survey in multiple formats. Managing this complex database requires significant organizational skills, attention to detail and ability to communicate within collective team members from diverse disciplines and across multiple institutions in the US and Mongolia. We describe the database's rich content, organization, structure and complexity. We discuss lessons learned, best practices and recommendations for complex database management, sharing, and archiving in creating a cross-cultural and multi-disciplinary database.

  16. Automated Fault Interpretation and Extraction using Improved Supplementary Seismic Datasets

    Science.gov (United States)

    Bollmann, T. A.; Shank, R.

    2017-12-01

    During the interpretation of seismic volumes, it is necessary to interpret faults along with horizons of interest. With the improvement of technology, the interpretation of faults can be expedited with the aid of different algorithms that create supplementary seismic attributes, such as semblance and coherency. These products highlight discontinuities, but still need a large amount of human interaction to interpret faults and are plagued by noise and stratigraphic discontinuities. Hale (2013) presents a method to improve on these datasets by creating what is referred to as a Fault Likelihood volume. In general, these volumes contain less noise and do not emphasize stratigraphic features. Instead, planar features within a specified strike and dip range are highlighted. Once a satisfactory Fault Likelihood Volume is created, extraction of fault surfaces is much easier. The extracted fault surfaces are then exported to interpretation software for QC. Numerous software packages have implemented this methodology with varying results. After investigating these platforms, we developed a preferred Automated Fault Interpretation workflow.

  17. Privacy-preserving record linkage on large real world datasets.

    Science.gov (United States)

    Randall, Sean M; Ferrante, Anna M; Boyd, James H; Bauer, Jacqueline K; Semmens, James B

    2014-08-01

    Record linkage typically involves the use of dedicated linkage units who are supplied with personally identifying information to determine individuals from within and across datasets. The personally identifying information supplied to linkage units is separated from clinical information prior to release by data custodians. While this substantially reduces the risk of disclosure of sensitive information, some residual risks still exist and remain a concern for some custodians. In this paper we trial a method of record linkage which reduces privacy risk still further on large real world administrative data. The method uses encrypted personal identifying information (bloom filters) in a probability-based linkage framework. The privacy preserving linkage method was tested on ten years of New South Wales (NSW) and Western Australian (WA) hospital admissions data, comprising in total over 26 million records. No difference in linkage quality was found when the results were compared to traditional probabilistic methods using full unencrypted personal identifiers. This presents as a possible means of reducing privacy risks related to record linkage in population level research studies. It is hoped that through adaptations of this method or similar privacy preserving methods, risks related to information disclosure can be reduced so that the benefits of linked research taking place can be fully realised. Copyright © 2013 Elsevier Inc. All rights reserved.

  18. Genomics dataset on unclassified published organism (patent US 7547531

    Directory of Open Access Journals (Sweden)

    Mohammad Mahfuz Ali Khan Shawan

    2016-12-01

    Full Text Available Nucleotide (DNA sequence analysis provides important clues regarding the characteristics and taxonomic position of an organism. With the intention that, DNA sequence analysis is very crucial to learn about hierarchical classification of that particular organism. This dataset (patent US 7547531 is chosen to simplify all the complex raw data buried in undisclosed DNA sequences which help to open doors for new collaborations. In this data, a total of 48 unidentified DNA sequences from patent US 7547531 were selected and their complete sequences were retrieved from NCBI BioSample database. Quick response (QR code of those DNA sequences was constructed by DNA BarID tool. QR code is useful for the identification and comparison of isolates with other organisms. AT/GC content of the DNA sequences was determined using ENDMEMO GC Content Calculator, which indicates their stability at different temperature. The highest GC content was observed in GP445188 (62.5% which was followed by GP445198 (61.8% and GP445189 (59.44%, while lowest was in GP445178 (24.39%. In addition, New England BioLabs (NEB database was used to identify cleavage code indicating the 5, 3 and blunt end and enzyme code indicating the methylation site of the DNA sequences was also shown. These data will be helpful for the construction of the organisms’ hierarchical classification, determination of their phylogenetic and taxonomic position and revelation of their molecular characteristics.

  19. Structural dataset for the PPARγ V290M mutant

    Directory of Open Access Journals (Sweden)

    Ana C. Puhl

    2016-06-01

    Full Text Available Loss-of-function mutation V290M in the ligand-binding domain of the peroxisome proliferator activated receptor γ (PPARγ is associated with a ligand resistance syndrome (PLRS, characterized by partial lipodystrophy and severe insulin resistance. In this data article we discuss an X-ray diffraction dataset that yielded the structure of PPARγ LBD V290M mutant refined at 2.3 Å resolution, that allowed building of 3D model of the receptor mutant with high confidence and revealed continuous well-defined electron density for the partial agonist diclofenac bound to hydrophobic pocket of the PPARγ. These structural data provide significant insights into molecular basis of PLRS caused by V290M mutation and are correlated with the receptor disability of rosiglitazone binding and increased affinity for corepressors. Furthermore, our structural evidence helps to explain clinical observations which point out to a failure to restore receptor function by the treatment with a full agonist of PPARγ, rosiglitazone.

  20. Storage flux uncertainty impact on eddy covariance net ecosystem exchange measurements

    Science.gov (United States)

    Nicolini, Giacomo; Aubinet, Marc; Feigenwinter, Christian; Heinesch, Bernard; Lindroth, Anders; Mamadou, Ossénatou; Moderow, Uta; Mölder, Meelis; Montagnani, Leonardo; Rebmann, Corinna; Papale, Dario

    2017-04-01

    Complying with several assumption and simplifications, most of the carbon budget studies based on eddy covariance (EC) measurements, quantify the net ecosystem exchange (NEE) by summing the flux obtained by EC (Fc) and the storage flux (Sc). Sc is the rate of change of CO2, within the so called control volume below the EC measurement level, given by the difference in the instantaneous profiles of concentration at the beginning and end of the EC averaging period, divided by the averaging period. While cumulating over time led to a nullification of Sc, it can be significant at short time periods. The approaches used to estimate Sc fluxes largely vary, from measurements based only on a single sampling point (usually located at the EC measurement height) to measurements based on several sampling profiles distributed within the control volume. Furthermore, the number of sampling points within each profile vary, according to their height and the ecosystem typology. It follows that measurement accuracy increases with the sampling intensity within the control volume. In this work we use the experimental dataset collected during the ADVEX campaign in which Sc flux has been measured in three similar forest sites by the use of 5 sampling profiles (towers). Our main objective is to quantify the impact of Sc measurement uncertainty on NEE estimates. Results show that different methods may produce substantially different Sc flux estimates, with problematic consequences in case high frequency (half-hourly) data are needed for the analysis. However, the uncertainty on long-term estimates may be tolerate.

  1. Oil palm mapping for Malaysia using PALSAR-2 dataset

    Science.gov (United States)

    Gong, P.; Qi, C. Y.; Yu, L.; Cracknell, A.

    2016-12-01

    Oil palm is one of the most productive vegetable oil crops in the world. The main oil palm producing areas are distributed in humid tropical areas such as Malaysia, Indonesia, Thailand, western and central Africa, northern South America, and central America. Increasing market demands, high yields and low production costs of palm oil are the primary factors driving large-scale commercial cultivation of oil palm, especially in Malaysia and Indonesia. Global demand for palm oil has grown exponentially during the last 50 years, and the expansion of oil palm plantations is linked directly to the deforestation of natural forests. Satellite remote sensing plays an important role in monitoring expansion of oil palm. However, optical remote sensing images are difficult to acquire in the Tropics because of the frequent occurrence of thick cloud cover. This problem has led to the use of data obtained by synthetic aperture radar (SAR), which is a sensor capable of all-day/all-weather observation for studies in the Tropics. In this study, the ALOS-2 (Advanced Land Observing Satellite) PALSAR-2 (Phased Array type L-band SAR) datasets for year 2015 were used as an input to a support vector machine (SVM) based machine learning algorithm. Oil palm/non-oil palm samples were collected using a hexagonal equal-area sampling design. High-resolution images in Google Earth and PALSAR-2 imagery were used in human photo-interpretation to separate oil palm from others (i.e. cropland, forest, grassland, shrubland, water, hard surface and bareland). The characteristics of oil palms from various aspects, including PALSAR-2 backscattering coefficients (HH, HV), terrain and climate by using this sample set were further explored to post-process the SVM output. The average accuracy of oil palm type is better than 80% in the final oil palm map for Malaysia.

  2. Automatic aortic root segmentation in CTA whole-body dataset

    Science.gov (United States)

    Gao, Xinpei; Kitslaar, Pieter H.; Scholte, Arthur J. H. A.; Lelieveldt, Boudewijn P. F.; Dijkstra, Jouke; Reiber, Johan H. C.

    2016-03-01

    Trans-catheter aortic valve replacement (TAVR) is an evolving technique for patients with serious aortic stenosis disease. Typically, in this application a CTA data set is obtained of the patient's arterial system from the subclavian artery to the femoral arteries, to evaluate the quality of the vascular access route and analyze the aortic root to determine if and which prosthesis should be used. In this paper, we concentrate on the automated segmentation of the aortic root. The purpose of this study was to automatically segment the aortic root in computed tomography angiography (CTA) datasets to support TAVR procedures. The method in this study includes 4 major steps. First, the patient's cardiac CTA image was resampled to reduce the computation time. Next, the cardiac CTA image was segmented using an atlas-based approach. The most similar atlas was selected from a total of 8 atlases based on its image similarity to the input CTA image. Third, the aortic root segmentation from the previous step was transferred to the patient's whole-body CTA image by affine registration and refined in the fourth step using a deformable subdivision surface model fitting procedure based on image intensity. The pipeline was applied to 20 patients. The ground truth was created by an analyst who semi-automatically corrected the contours of the automatic method, where necessary. The average Dice similarity index between the segmentations of the automatic method and the ground truth was found to be 0.965±0.024. In conclusion, the current results are very promising.

  3. Local rectification of heat flux

    Science.gov (United States)

    Pons, M.; Cui, Y. Y.; Ruschhaupt, A.; Simón, M. A.; Muga, J. G.

    2017-09-01

    We present a chain-of-atoms model where heat is rectified, with different fluxes from the hot to the cold baths located at the chain boundaries when the temperature bias is reversed. The chain is homogeneous except for boundary effects and a local modification of the interactions at one site, the “impurity”. The rectification mechanism is due here to the localized impurity, the only asymmetrical element of the structure, apart from the externally imposed temperature bias, and does not rely on putting in contact different materials or other known mechanisms such as grading or long-range interactions. The effect survives if all interaction forces are linear except the ones for the impurity.

  4. LOFT gamma densitometer background fluxes

    International Nuclear Information System (INIS)

    Grimesey, R.A.; McCracken, R.T.

    1978-01-01

    Background gamma-ray fluxes were calculated at the location of the γ densitometers without integral shielding at both the hot-leg and cold-leg primary piping locations. The principal sources for background radiation at the γ densitometers are 16 N activity from the primary piping H 2 O and γ radiation from reactor internal sources. The background radiation was calculated by the point-kernel codes QAD-BSA and QAD-P5A. Reasonable assumptions were required to convert the response functions calculated by point-kernel procedures into the gamma-ray spectrum from reactor internal sources. A brief summary of point-kernel equations and theory is included

  5. Nuclear transmutation by flux compression

    International Nuclear Information System (INIS)

    Seifritz, W.

    2001-01-01

    A new idea for the transmutation of minor actinides, long (and even short) lived fission products is presented. It is based an the property of neutron flux compression in nuclear (fast and/or thermal) reactors possessing spatially non-stationary critical masses. An advantage factor for the burn-up fluence of the elements to be transmuted in the order of magnitude of 100 and more is obtainable compared with the classical way of transmutation. Three typical examples of such transmuters (a subcritical ringreactor with a rotating reflector, a sub-critical ring reactor with a rotating spallation source, the socalled ''pulsed energy amplifier'', and a fast burn-wave reactor) are presented and analysed with regard to this purpose. (orig.) [de

  6. Dynamics of warped flux compactifications

    International Nuclear Information System (INIS)

    Shiu, Gary; Underwood, Bret; Torroba, Gonzalo; Douglas, Michael R.

    2008-01-01

    We discuss the four dimensional effective action for type IIB flux compactifications, and obtain the quadratic terms taking warp effects into account. The analysis includes both the 4-d zero modes and their KK excitations, which become light at large warping. We identify an 'axial' type gauge for the supergravity fluctuations, which makes the four dimensional degrees of freedom manifest. The other key ingredient is the existence of constraints coming from the ten dimensional equations of motion. Applying these conditions leads to considerable simplifications, enabling us to obtain the low energy lagrangian explicitly. In particular, the warped Kaehler potential for metric moduli is computed and it is shown that there are no mixings with the KK fluctuations and the result differs from previous proposals. The four dimensional potential contains a generalization of the Gukov-Vafa-Witten term, plus usual mass terms for KK modes.

  7. Large biases in regression-based constituent flux estimates: causes and diagnostic tools

    Science.gov (United States)

    Hirsch, Robert M.

    2014-01-01

    It has been documented in the literature that, in some cases, widely used regression-based models can produce severely biased estimates of long-term mean river fluxes of various constituents. These models, estimated using sample values of concentration, discharge, and date, are used to compute estimated fluxes for a multiyear period at a daily time step. This study compares results of the LOADEST seven-parameter model, LOADEST five-parameter model, and the Weighted Regressions on Time, Discharge, and Season (WRTDS) model using subsampling of six very large datasets to better understand this bias problem. This analysis considers sample datasets for dissolved nitrate and total phosphorus. The results show that LOADEST-7 and LOADEST-5, although they often produce very nearly unbiased results, can produce highly biased results. This study identifies three conditions that can give rise to these severe biases: (1) lack of fit of the log of concentration vs. log discharge relationship, (2) substantial differences in the shape of this relationship across seasons, and (3) severely heteroscedastic residuals. The WRTDS model is more resistant to the bias problem than the LOADEST models but is not immune to them. Understanding the causes of the bias problem is crucial to selecting an appropriate method for flux computations. Diagnostic tools for identifying the potential for bias problems are introduced, and strategies for resolving bias problems are described.

  8. Pyrolytic graphite gauge for measuring heat flux

    Science.gov (United States)

    Bunker, Robert C. (Inventor); Ewing, Mark E. (Inventor); Shipley, John L. (Inventor)

    2002-01-01

    A gauge for measuring heat flux, especially heat flux encountered in a high temperature environment, is provided. The gauge includes at least one thermocouple and an anisotropic pyrolytic graphite body that covers at least part of, and optionally encases the thermocouple. Heat flux is incident on the anisotropic pyrolytic graphite body by arranging the gauge so that the gauge surface on which convective and radiative fluxes are incident is perpendicular to the basal planes of the pyrolytic graphite. The conductivity of the pyrolytic graphite permits energy, transferred into the pyrolytic graphite body in the form of heat flux on the incident (or facing) surface, to be quickly distributed through the entire pyrolytic graphite body, resulting in small substantially instantaneous temperature gradients. Temperature changes to the body can thereby be measured by the thermocouple, and reduced to quantify the heat flux incident to the body.

  9. Minkowski vacuum transitions in (nongeometric) flux compactifications

    International Nuclear Information System (INIS)

    Herrera-Suarez, Wilberth; Loaiza-Brito, Oscar

    2010-01-01

    In this work we study the generalization of twisted homology to geometric and nongeometric backgrounds. In the process, we describe the necessary conditions to wrap a network of D-branes on twisted cycles. If the cycle is localized in time, we show how by an instantonic brane mediation, some D-branes transform into fluxes on different backgrounds, including nongeometric fluxes. As a consequence, we show that in the case of a IIB six-dimensional torus compactification on a simple orientifold, the flux superpotential is not invariant by this brane-flux transition, allowing the connection among different Minkowski vacuum solutions. For the case in which nongeometric fluxes are turned on, we also discuss some topological restrictions for the transition to occur. In this context, we show that there are some vacuum solutions protected to change by a brane-flux transition.

  10. rf SQUID system as tunable flux qubit

    Energy Technology Data Exchange (ETDEWEB)

    Ruggiero, B. [Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy)]. E-mail: b.ruggiero@cib.na.cnr.it; Granata, C. [Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy); Vettoliere, A. [Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy); Rombetto, S. [Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy); Russo, R. [Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy); Russo, M. [Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy); Corato, V. [Dipartimento di Ingegneria dell' Informazione, Seconda Universita di Napoli, I-81031 Aversa (Italy); Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy); Silvestrini, P. [Dipartimento di Ingegneria dell' Informazione, Seconda Universita di Napoli, I-81031 Aversa (Italy); Istituto di Cibernetica ' E. Caianiello' del Consiglio Nazionale delle Ricerche, I-80078 Pozzuoli (Italy)

    2006-08-21

    We present a fully integrated rf SQUID-based system as flux qubit with a high control of the flux transfer function of the superconducting transformer modulating the coupling between the flux qubit and the readout system. The control of the system is possible by including into the superconducting flux transformer a vertical two-Josephson-junctions interferometer (VJI) in which the Josephson current is precisely modulated from a maximum to zero by a transversal magnetic field parallel to the flux transformer plane. The proposed system can be also used in a more general configuration to control the off-diagonal terms in the Hamiltonian of the flux qubit and to turn on and off the coupling between two or more qubits.

  11. Introduction of a simple-model-based land surface dataset for Europe

    Science.gov (United States)

    Orth, Rene; Seneviratne, Sonia I.

    2015-04-01

    Land surface hydrology can play a crucial role during extreme events such as droughts, floods and even heat waves. We introduce in this study a new hydrological dataset for Europe that consists of soil moisture, runoff and evapotranspiration (ET). It is derived with a simple water balance model (SWBM) forced with precipitation, temperature and net radiation. The SWBM dataset extends over the period 1984-2013 with a daily time step and 0.5° × 0.5° resolution. We employ a novel calibration approach, in which we consider 300 random parameter sets chosen from an observation-based range. Using several independent validation datasets representing soil moisture (or terrestrial water content), ET and streamflow, we identify the best performing parameter set and hence the new dataset. To illustrate its usefulness, the SWBM dataset is compared against several state-of-the-art datasets (ERA-Interim/Land, MERRA-Land, GLDAS-2-Noah, simulations of the Community Land Model Version 4), using all validation datasets as reference. For soil moisture dynamics it outperforms the benchmarks. Therefore the SWBM soil moisture dataset constitutes a reasonable alternative to sparse measurements, little validated model results, or proxy data such as precipitation indices. Also in terms of runoff the SWBM dataset performs well, whereas the evaluation of the SWBM ET dataset is overall satisfactory, but the dynamics are less well captured for this variable. This highlights the limitations of the dataset, as it is based on a simple model that uses uniform parameter values. Hence some processes impacting ET dynamics may not be captured, and quality issues may occur in regions with complex terrain. Even though the SWBM is well calibrated, it cannot replace more sophisticated models; but as their calibration is a complex task the present dataset may serve as a benchmark in future. In addition we investigate the sources of skill of the SWBM dataset and find that the parameter set has a similar

  12. Fast heat flux modulation at the nanoscale

    OpenAIRE

    van Zwol, P. J.; Joulain, K.; Abdallah, P. Ben; Greffet, J. J.; Chevrier, J.

    2011-01-01

    We introduce a new concept for electrically controlled heat flux modulation. A flux contrast larger than 10 dB is expected with switching time on the order of tens of nanoseconds. Heat flux modulation is based on the interplay between radiative heat transfer at the nanoscale and phase change materials. Such large contrasts are not obtainable in solids, or in far field. As such this opens up new horizons for temperature modulation and actuation at the nanoscale.

  13. Heat Flux Inhibition by Whistlers: Experimental Confirmation

    International Nuclear Information System (INIS)

    Eichler, D.

    2002-01-01

    Heat flux in weakly magnetized collisionless plasma is, according to theoretical predictions, limited by whistler turbulence that is generated by heat flux instabilities near threshold. Observations of solar wind electrons by Gary and coworkers appear to confirm the limit on heat flux as being roughly the product of the magnetic energy density and the electron thermal velocity, in agreement with prediction (Pistinner and Eichler 1998)

  14. VideoWeb Dataset for Multi-camera Activities and Non-verbal Communication

    Science.gov (United States)

    Denina, Giovanni; Bhanu, Bir; Nguyen, Hoang Thanh; Ding, Chong; Kamal, Ahmed; Ravishankar, Chinya; Roy-Chowdhury, Amit; Ivers, Allen; Varda, Brenda

    Human-activity recognition is one of the most challenging problems in computer vision. Researchers from around the world have tried to solve this problem and have come a long way in recognizing simple motions and atomic activities. As the computer vision community heads toward fully recognizing human activities, a challenging and labeled dataset is needed. To respond to that need, we collected a dataset of realistic scenarios in a multi-camera network environment (VideoWeb) involving multiple persons performing dozens of different repetitive and non-repetitive activities. This chapter describes the details of the dataset. We believe that this VideoWeb Activities dataset is unique and it is one of the most challenging datasets available today. The dataset is publicly available online at http://vwdata.ee.ucr.edu/ along with the data annotation.

  15. Dimensional reduction of a generalized flux problem

    International Nuclear Information System (INIS)

    Moroz, A.

    1992-01-01

    In this paper, a generalized flux problem with Abelian and non-Abelian fluxes is considered. In the Abelian case we shall show that the generalized flux problem for tight-binding models of noninteracting electrons on either 2n- or (2n + 1)-dimensional lattice can always be reduced to an n-dimensional hopping problem. A residual freedom in this reduction enables one to identify equivalence classes of hopping Hamiltonians which have the same spectrum. In the non-Abelian case, the reduction is not possible in general unless the flux tensor factorizes into an Abelian one times are element of the corresponding algebra

  16. USGS Watershed Boundary Dataset (WBD) Overlay Map Service from The National Map - National Geospatial Data Asset (NGDA) Watershed Boundary Dataset (WBD)

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — The Watershed Boundary Dataset (WBD) from The National Map (TNM) defines the perimeter of drainage areas formed by the terrain and other landscape characteristics....

  17. The StreamCat Dataset: Accumulated Attributes for NHDPlusV2 Catchments (Version 2.1) for the Conterminous United States: National Coal Resource Dataset System

    Data.gov (United States)

    U.S. Environmental Protection Agency — This dataset represents the coal mine density and storage volumes within individual, local NHDPlusV2 catchments and upstream, contributing watersheds based on the...

  18. The StreamCat Dataset: Accumulated Attributes for NHDPlusV2 (Version 2.1) Catchments for the Conterminous United States: National Anthropogenic Barrier Dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — This dataset represents the dam density and storage volumes within individual, local NHDPlusV2 catchments and upstream, contributing watersheds based on the National...

  19. The StreamCat Dataset: Accumulated Attributes for NHDPlusV2 (Version 2.1) Catchments for the Conterminous United States: National Elevation Dataset

    Data.gov (United States)

    U.S. Environmental Protection Agency — This dataset represents the elevation values within individual local NHDPlusV2 catchments and upstream, contributing watersheds based on the National Elevation...

  20. Short-term impacts of enhanced Greenland freshwater fluxes in an eddy-permitting ocean model

    Directory of Open Access Journals (Sweden)

    R. Marsh

    2010-07-01

    Full Text Available In a sensitivity experiment, an eddy-permitting ocean general circulation model is forced with realistic freshwater fluxes from the Greenland Ice Sheet, averaged for the period 1991–2000. The fluxes are obtained with a mass balance model for the ice sheet, forced with the ERA-40 reanalysis dataset. The freshwater flux is distributed around Greenland as an additional term in prescribed runoff, representing seasonal melting of the ice sheet and a fixed year-round iceberg calving flux, for 8.5 model years. By adding Greenland freshwater fluxes with realistic geographical distribution and seasonality, the experiment is designed to investigate the oceanic response to a sudden and spatially/temporally uniform amplification of ice sheet melting and discharge, rather than localized or gradual changes in freshwater flux. The impacts on regional hydrography and circulation are investigated by comparing the sensitivity experiment to a control experiment, without additional fluxes. By the end of the sensitivity experiment, the majority of additional fresh water has accumulated in Baffin Bay, and only a small fraction has reached the interior of the Labrador Sea, where winter mixed layer depth is sensitive to small changes in salinity. As a consequence, the impact on large-scale circulation is very slight. An indirect impact of strong freshening off the west coast of Greenland is a small anti-cyclonic component to the circulation around Greenland, which opposes the wind-driven cyclonic circulation and reduces net southward flow through the Canadian Archipelago by ~10%. Implications for the post-2000 acceleration of Greenland mass loss are discussed.

  1. Topic modeling for cluster analysis of large biological and medical datasets.

    Science.gov (United States)

    Zhao, Weizhong; Zou, Wen; Chen, James J

    2014-01-01

    The big data moniker is nowhere better deserved than to describe the ever-increasing prodigiousness and complexity of biological and medical datasets. New methods are needed to generate and test hypotheses, foster biological interpretation, and build validated predictors. Although multivariate techniques such as cluster analysis may allow researchers to identify groups, or clusters, of related variables, the accuracies and effectiveness of traditional clustering methods diminish for large and hyper dimensional datasets. Topic modeling is an active research field in machine learning and has been mainly used as an analytical tool to structure large textual corpora for data mining. Its ability to reduce high dimensionality to a small number of latent variables makes it suitable as a means for clustering or overcoming clustering difficulties in large biological and medical datasets. In this study, three topic model-derived clustering methods, highest probable topic assignment, feature selection and feature extraction, are proposed and tested on the cluster analysis of three large datasets: Salmonella pulsed-field gel electrophoresis (PFGE) dataset, lung cancer dataset, and breast cancer dataset, which represent various types of large biological or medical datasets. All three various methods are shown to improve the efficacy/effectiveness of clustering results on the three datasets in comparison to traditional methods. A preferable cluster analysis method emerged for each of the three datasets on the basis of replicating known biological truths. Topic modeling could be advantageously applied to the large datasets of biological or medical research. The three proposed topic model-derived clustering methods, highest probable topic assignment, feature selection and feature extraction, yield clustering improvements for the three different data types. Clusters more efficaciously represent truthful groupings and subgroupings in the data than traditional methods, suggesting

  2. A new dataset and algorithm evaluation for mood estimation in music

    OpenAIRE

    Godec, Primož

    2014-01-01

    This thesis presents a new dataset of perceived and induced emotions for 200 audio clips. The gathered dataset provides users' perceived and induced emotions for each clip, the association of color, along with demographic and personal data, such as user's emotion state and emotion ratings, genre preference, music experience, among others. With an online survey we collected more than 7000 responses for a dataset of 200 audio excerpts, thus providing about 37 user responses per clip. The foc...

  3. Relating Radiative Fluxes on Arctic Sea Ice Area Using Arctic Observation and Reanalysis Integrated System (ArORIS)

    Science.gov (United States)

    Sledd, A.; L'Ecuyer, T. S.

    2017-12-01

    With Arctic sea ice declining rapidly and Arctic temperatures rising faster than the rest of the globe, a better understanding of the Arctic climate, and ice cover-radiation feedbacks in particular, is needed. Here we present the Arctic Observation and Reanalysis Integrated System (ArORIS), a dataset of integrated products to facilitate studying the Arctic using satellite, reanalysis, and in-situ datasets. The data include cloud properties, radiative fluxes, aerosols, meteorology, precipitation, and surface properties, to name just a few. Each dataset has uniform grid-spacing, time-averaging and naming conventions for ease of use between products. One intended use of ArORIS is to assess Arctic radiation and moisture budgets. Following that goal, we use observations from ArORIS - CERES-EBAF radiative fluxes and NSIDC sea ice fraction and area to quantify relationships between the Arctic energy balance and surface properties. We find a discernable difference between energy budgets for years with high and low September sea ice areas. Surface fluxes are especially responsive to the September sea ice minimum in months both leading up to September and the months following. In particular, longwave fluxes at the surface show increased sensitivity in the months preceding September. Using a single-layer model of solar radiation we also investigate the individual responses of surface and planetary albedos to changes in sea ice area. By partitioning the planetary albedo into surface and atmospheric contributions, we find that the atmospheric contribution to planetary albedo is less sensitive to changes in sea ice area than the surface contribution. Further comparisons between observations and reanalyses can be made using the available datasets in ArORIS.

  4. Hydrological simulation of the Brahmaputra basin using global datasets

    Science.gov (United States)

    Bhattacharya, Biswa; Conway, Crystal; Craven, Joanne; Masih, Ilyas; Mazzolini, Maurizio; Shrestha, Shreedeepy; Ugay, Reyne; van Andel, Schalk Jan

    2017-04-01

    Brahmaputra River flows through China, India and Bangladesh to the Bay of Bengal and is one of the largest rivers of the world with a catchment size of 580K km2. The catchment is largely hilly and/or forested with sparse population and with limited urbanisation and economic activities. The catchment experiences heavy monsoon rainfall leading to very high flood discharges. Large inter-annual variation of discharge leading to flooding, erosion and morphological changes are among the major challenges. The catchment is largely ungauged; moreover, limited availability of hydro-meteorological data limits the possibility of carrying out evidence based research, which could provide trustworthy information for managing and when needed, controlling, the basin processes by the riparian countries for overall basin development. The paper presents initial results of a current research project on Brahmaputra basin. A set of hydrological and hydraulic models (SWAT, HMS, RAS) are developed by employing publicly available datasets of DEM, land use and soil and simulated using satellite based rainfall products, evapotranspiration and temperature estimates. Remotely sensed data are compared with sporadically available ground data. The set of models are able to produce catchment wide hydrological information that potentially can be used in the future in managing the basin's water resources. The model predications should be used with caution due to high level of uncertainty because the semi-calibrated models are developed with uncertain physical representation (e.g. cross-section) and simulated with global meteorological forcing (e.g. TRMM) with limited validation. Major scientific challenges are seen in producing robust information that can be reliably used in managing the basin. The information generated by the models are uncertain and as a result, instead of using them per se, they are used in improving the understanding of the catchment, and by running several scenarios with varying

  5. Investigating automated depth modelling of archaeo-magnetic datasets

    Science.gov (United States)

    Cheyney, Samuel; Hill, Ian; Linford, Neil; Leech, Christopher

    2010-05-01

    Magnetic surveying is a commonly used tool for first-pass non-invasive archaeological surveying, and is often used to target areas for more detailed geophysical investigation, or excavation. Quick and routine processing of magnetic datasets mean survey results are typically viewed as 2D greyscale maps and the shapes of anomalies are interpreted in terms of likely archaeological structures. This technique is simple, but ignores some of the information content of the data. The data collected using dense spatial sampling with modern precise instrumentation are capable of yielding numerical estimates of the depths to buried structures, and their physical properties. The magnetic field measured at the surface is a superposition of the responses to all anomalous magnetic susceptibilities in the subsurface, and is therefore capable of revealing a 3D model of the magnetic properties. The application of mathematical modelling techniques to very-near-surface surveys such as for archaeology is quite rare, however similar methods are routinely used in regional scale mineral exploration surveys. Inverse modelling techniques have inherent ambiguity due to the nature of the mathematical "inverse problem". Often, although a good fit to the recorded values can be obtained, the final model will be non-unique and may be heavily biased by the starting model provided. Also the run time and computer resources required can be restrictive. Our approach is to derive as much information as possible from the data directly, and use this to define a starting model for inversion. This addresses both the ambiguity of the inverse problem and reduces the task for the inversion computation. A number of alternative methods exist that can be used to obtain parameters for source bodies in potential field data. Here, methods involving the derivatives of the total magnetic field are used in association with advanced image processing techniques to outline the edges of anomalous bodies more accurately

  6. Reliability of Source Mechanisms for a Hydraulic Fracturing Dataset

    Science.gov (United States)

    Eyre, T.; Van der Baan, M.

    2016-12-01

    Non-double-couple components have been inferred for induced seismicity due to fluid injection, yet these components are often poorly constrained due to the acquisition geometry. Likewise non-double-couple components in microseismic recordings are not uncommon. Microseismic source mechanisms provide an insight into the fracturing behaviour of a hydraulically stimulated reservoir. However, source inversion in a hydraulic fracturing environment is complicated by the likelihood of volumetric contributions to the source due to the presence of high pressure fluids, which greatly increases the possible solution space and therefore the non-uniqueness of the solutions. Microseismic data is usually recorded on either 2D surface or borehole arrays of sensors. In many cases, surface arrays appear to constrain source mechanisms with high shear components, whereas borehole arrays tend to constrain more variable mechanisms including those with high tensile components. The abilities of each geometry to constrain the true source mechanisms are therefore called into question.The ability to distinguish between shear and tensile source mechanisms with different acquisition geometries is investigated using synthetic data. For both inversions, both P- and S- wave amplitudes recorded on three component sensors need to be included to obtain reliable solutions. Surface arrays appear to give more reliable solutions due to a greater sampling of the focal sphere, but in reality tend to record signals with a low signal to noise ratio. Borehole arrays can produce acceptable results, however the reliability is much more affected by relative source-receiver locations and source orientation, with biases produced in many of the solutions. Therefore more care must be taken when interpreting results.These findings are taken into account when interpreting a microseismic dataset of 470 events recorded by two vertical borehole arrays monitoring a horizontal treatment well. Source locations and

  7. Estimated Perennial Streams of Idaho and Related Geospatial Datasets

    Science.gov (United States)

    Rea, Alan; Skinner, Kenneth D.

    2009-01-01

    record, generally would be considered to represent flow conditions better at a given site than flow estimates based on regionalized regression models. The geospatial datasets of modeled perennial streams are considered a first-cut estimate, and should not be construed to override site-specific flow data.

  8. Scalable and portable visualization of large atomistic datasets

    Science.gov (United States)

    Sharma, Ashish; Kalia, Rajiv K.; Nakano, Aiichiro; Vashishta, Priya

    2004-10-01

    A scalable and portable code named Atomsviewer has been developed to interactively visualize a large atomistic dataset consisting of up to a billion atoms. The code uses a hierarchical view frustum-culling algorithm based on the octree data structure to efficiently remove atoms outside of the user's field-of-view. Probabilistic and depth-based occlusion-culling algorithms then select atoms, which have a high probability of being visible. Finally a multiresolution algorithm is used to render the selected subset of visible atoms at varying levels of detail. Atomsviewer is written in C++ and OpenGL, and it has been tested on a number of architectures including Windows, Macintosh, and SGI. Atomsviewer has been used to visualize tens of millions of atoms on a standard desktop computer and, in its parallel version, up to a billion atoms. Program summaryTitle of program: Atomsviewer Catalogue identifier: ADUM Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADUM Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Computer for which the program is designed and others on which it has been tested: 2.4 GHz Pentium 4/Xeon processor, professional graphics card; Apple G4 (867 MHz)/G5, professional graphics card Operating systems under which the program has been tested: Windows 2000/XP, Mac OS 10.2/10.3, SGI IRIX 6.5 Programming languages used: C++, C and OpenGL Memory required to execute with typical data: 1 gigabyte of RAM High speed storage required: 60 gigabytes No. of lines in the distributed program including test data, etc.: 550 241 No. of bytes in the distributed program including test data, etc.: 6 258 245 Number of bits in a word: Arbitrary Number of processors used: 1 Has the code been vectorized or parallelized: No Distribution format: tar gzip file Nature of physical problem: Scientific visualization of atomic systems Method of solution: Rendering of atoms using computer graphic techniques, culling algorithms for data

  9. GUDM: Automatic Generation of Unified Datasets for Learning and Reasoning in Healthcare.

    Science.gov (United States)

    Ali, Rahman; Siddiqi, Muhammad Hameed; Idris, Muhammad; Ali, Taqdir; Hussain, Shujaat; Huh, Eui-Nam; Kang, Byeong Ho; Lee, Sungyoung

    2015-07-02

    A wide array of biomedical data are generated and made available to healthcare experts. However, due to the diverse nature of data, it is difficult to predict outcomes from it. It is therefore necessary to combine these diverse data sources into a single unified dataset. This paper proposes a global unified data model (GUDM) to provide a global unified data structure for all data sources and generate a unified dataset by a "data modeler" tool. The proposed tool implements user-centric priority based approach which can easily resolve the problems of unified data modeling and overlapping attributes across multiple datasets. The tool is illustrated using sample diabetes mellitus data. The diverse data sources to generate the unified dataset for diabetes mellitus include clinical trial information, a social media interaction dataset and physical activity data collected using different sensors. To realize the significance of the unified dataset, we adopted a well-known rough set theory based rules creation process to create rules from the unified dataset. The evaluation of the tool on six different sets of locally created diverse datasets shows that the tool, on average, reduces 94.1% time efforts of the experts and knowledge engineer while creating unified datasets.

  10. Wehmas et al. 94-04 Toxicol Sci: Datasets for manuscript

    Data.gov (United States)

    U.S. Environmental Protection Agency — Dataset includes overview text document (accepted version of manuscript) and tables, figures, and supplementary materials. Supplementary tables provide summary data...

  11. A New Dataset Size Reduction Approach for PCA-Based Classification in OCR Application

    Directory of Open Access Journals (Sweden)

    Mohammad Amin Shayegan

    2014-01-01

    Full Text Available A major problem of pattern recognition systems is due to the large volume of training datasets including duplicate and similar training samples. In order to overcome this problem, some dataset size reduction and also dimensionality reduction techniques have been introduced. The algorithms presently used for dataset size reduction usually remove samples near to the centers of classes or support vector samples between different classes. However, the samples near to a class center include valuable information about the class characteristics and the support vector is important for evaluating system efficiency. This paper reports on the use of Modified Frequency Diagram technique for dataset size reduction. In this new proposed technique, a training dataset is rearranged and then sieved. The sieved training dataset along with automatic feature extraction/selection operation using Principal Component Analysis is used in an OCR application. The experimental results obtained when using the proposed system on one of the biggest handwritten Farsi/Arabic numeral standard OCR datasets, Hoda, show about 97% accuracy in the recognition rate. The recognition speed increased by 2.28 times, while the accuracy decreased only by 0.7%, when a sieved version of the dataset, which is only as half as the size of the initial training dataset, was used.

  12. Topographical effects of climate dataset and their impacts on the estimation of regional net primary productivity

    Science.gov (United States)

    Sun, L. Qing; Feng, Feng X.

    2014-11-01

    In this study, we first built and compared two different climate datasets for Wuling mountainous area in 2010, one of which considered topographical effects during the ANUSPLIN interpolation was referred as terrain-based climate dataset, while the other one did not was called ordinary climate dataset. Then, we quantified the topographical effects of climatic inputs on NPP estimation by inputting two different climate datasets to the same ecosystem model, the Boreal Ecosystem Productivity Simulator (BEPS), to evaluate the importance of considering relief when estimating NPP. Finally, we found the primary contributing variables to the topographical effects through a series of experiments given an overall accuracy of the model output for NPP. The results showed that: (1) The terrain-based climate dataset presented more reliable topographic information and had closer agreements with the station dataset than the ordinary climate dataset at successive time series of 365 days in terms of the daily mean values. (2) On average, ordinary climate dataset underestimated NPP by 12.5% compared with terrain-based climate dataset over the whole study area. (3) The primary climate variables contributing to the topographical effects of climatic inputs for Wuling mountainous area were temperatures, which suggest that it is necessary to correct temperature differences for estimating NPP accurately in such a complex terrain.

  13. Nitrous Oxide flux measurements under various amendments

    Data.gov (United States)

    U.S. Environmental Protection Agency — The dataset consists of measurements of soil nitrous oxide emissions from soils under three different amendments: glucose, cellulose, and manure. Data includes the...

  14. Surface Flux Modeling for Air Quality Applications

    Directory of Open Access Journals (Sweden)

    Limei Ran

    2011-08-01

    Full Text Available For many gasses and aerosols, dry deposition is an important sink of atmospheric mass. Dry deposition fluxes are also important sources of pollutants to terrestrial and aquatic ecosystems. The surface fluxes of some gases, such as ammonia, mercury, and certain volatile organic compounds, can be upward into the air as well as downward to the surface and therefore should be modeled as bi-directional fluxes. Model parameterizations of dry deposition in air quality models have been represented by simple electrical resistance analogs for almost 30 years. Uncertainties in surface flux modeling in global to mesoscale models are being slowly reduced as more field measurements provide constraints on parameterizations. However, at the same time, more chemical species are being added to surface flux models as air quality models are expanded to include more complex chemistry and are being applied to a wider array of environmental issues. Since surface flux measurements of many of these chemicals are still lacking, resistances are usually parameterized using simple scaling by water or lipid solubility and reactivity. Advances in recent years have included bi-directional flux algorithms that require a shift from pre-computation of deposition velocities to fully integrated surface flux calculations within air quality models. Improved modeling of the stomatal component of chemical surface fluxes has resulted from improved evapotranspiration modeling in land surface models and closer integration between meteorology and air quality models. Satellite-derived land use characterization and vegetation products and indices are improving model representation of spatial and temporal variations in surface flux processes. This review describes the current state of chemical dry deposition modeling, recent progress in bi-directional flux modeling, synergistic model development research with field measurements, and coupling with meteorological land surface models.

  15. Magnetic flux generator for balanced membrane loudspeaker

    DEFF Research Database (Denmark)

    Rehder, Jörg; Rombach, Pirmin; Hansen, Ole

    2002-01-01

    This paper reports the development of a magnetic flux generator with an application in a hearing aid loudspeaker produced in microsystem technology (MST). The technology plans for two different designs for the magnetic flux generator utilizing a softmagnetic substrate or electroplated Ni...

  16. EL-2 reactor: Thermal neutron flux distribution

    International Nuclear Information System (INIS)

    Rousseau, A.; Genthon, J.P.

    1958-01-01

    The flux distribution of thermal neutrons in EL-2 reactor is studied. The reactor core and lattices are described as well as the experimental reactor facilities, in particular, the experimental channels and special facilities. The measurement shows that the thermal neutron flux increases in the central channel when enriched uranium is used in place of natural uranium. However the thermal neutron flux is not perturbed in the other reactor channels by the fuel modification. The macroscopic flux distribution is measured according the radial positioning of fuel rods. The longitudinal neutron flux distribution in a fuel rod is also measured and shows no difference between enriched and natural uranium fuel rods. In addition, measurements of the flux distribution have been effectuated for rods containing other material as steel or aluminium. The neutron flux distribution is also studied in all the experimental channels as well as in the thermal column. The determination of the distribution of the thermal neutron flux in all experimental facilities, the thermal column and the fuel channels has been made with a heavy water level of 1825 mm and is given for an operating power of 1000 kW. (M.P.)

  17. Neutron flux measurement by mobile detectors

    International Nuclear Information System (INIS)

    Verchain, M.

    1987-01-01

    Various incore instrumentation systems and their technological evolution are first reviewed. Then, for 1300 MWe PWR nuclear power plant, temperature and neutron flux measurement are described. Mobile fission chambers, with their large measuring range and accurate location allow a good knowledge of the core. Other incore measures are possible because of flux detector thimble tubes inserted in the reactor core [fr

  18. Anthropogenic heat flux estimation from space

    NARCIS (Netherlands)

    Chrysoulakis, Nektarios; Marconcini, Mattia; Gastellu-Etchegorry, Jean Philippe; Grimmond, C.S.B.; Feigenwinter, Christian; Lindberg, Fredrik; Frate, Del Fabio; Klostermann, Judith; Mitraka, Zina; Esch, Thomas; Landier, Lucas; Gabey, Andy; Parlow, Eberhard; Olofson, Frans

    2016-01-01

    H2020-Space project URBANFLUXES (URBan ANthrpogenic heat FLUX from Earth observation Satellites) investigates the potential of Copernicus Sentinels to retrieve anthropogenic heat flux, as a key component of the Urban Energy Budget (UEB). URBANFLUXES advances the current knowledge of the impacts

  19. ANthropogenic heat FLUX estimation from Space

    NARCIS (Netherlands)

    Chrysoulakis, Nektarios; Marconcini, Mattia; Gastellu-Etchegorry, Jean Philippe; Grimmong, C.S.B.; Feigenwinter, Christian; Lindberg, Fredrik; Frate, Del Fabio; Klostermann, Judith; Mi, Zina; Esch, Thomas; Landier, Lucas; Gabey, Andy; Parlow, Eberhard; Olofson, Frans

    2017-01-01

    The H2020-Space project URBANFLUXES (URBan ANthrpogenic heat FLUX from Earth observation Satellites) investigates the potential of Copernicus Sentinels to retrieve anthropogenic heat flux, as a key component of the Urban Energy Budget (UEB). URBANFLUXES advances the current knowledge of the

  20. Fast flux module detection using matroid theory.

    Science.gov (United States)

    Reimers, Arne C; Bruggeman, Frank J; Olivier, Brett G; Stougie, Leen

    2015-05-01

    Flux balance analysis (FBA) is one of the most often applied methods on genome-scale metabolic networks. Although FBA uniquely determines the optimal yield, the pathway that achieves this is usually not unique. The analysis of the optimal-yield flux space has been an open challenge. Flux variability analysis is only capturing some properties of the flux space, while elementary mode analysis is intractable due to the enormous number of elementary modes. However, it has been found by Kelk et al. (2012) that the space of optimal-yield fluxes decomposes into flux modules. These decompositions allow a much easier but still comprehensive analysis of the optimal-yield flux space. Using the mathematical definition of module introduced by Müller and Bockmayr (2013b), we discovered useful connections to matroid theory, through which efficient algorithms enable us to compute the decomposition into modules in a few seconds for genome-scale networks. Using that every module can be represented by one reaction that represents its function, in this article, we also present a method that uses this decomposition to visualize the interplay of modules. We expect the new method to replace flux variability analysis in the pipelines for metabolic networks.

  1. FILAMENT INTERACTION MODELED BY FLUX ROPE RECONNECTION

    International Nuclear Information System (INIS)

    Toeroek, T.; Chandra, R.; Pariat, E.; Demoulin, P.; Schmieder, B.; Aulanier, G.; Linton, M. G.; Mandrini, C. H.

    2011-01-01

    Hα observations of solar active region NOAA 10501 on 2003 November 20 revealed a very uncommon dynamic process: during the development of a nearby flare, two adjacent elongated filaments approached each other, merged at their middle sections, and separated again, thereby forming stable configurations with new footpoint connections. The observed dynamic pattern is indicative of 'slingshot' reconnection between two magnetic flux ropes. We test this scenario by means of a three-dimensional zero β magnetohydrodynamic simulation, using a modified version of the coronal flux rope model by Titov and Demoulin as the initial condition for the magnetic field. To this end, a configuration is constructed that contains two flux ropes which are oriented side-by-side and are embedded in an ambient potential field. The choice of the magnetic orientation of the flux ropes and of the topology of the potential field is guided by the observations. Quasi-static boundary flows are then imposed to bring the middle sections of the flux ropes into contact. After sufficient driving, the ropes reconnect and two new flux ropes are formed, which now connect the former adjacent flux rope footpoints of opposite polarity. The corresponding evolution of filament material is modeled by calculating the positions of field line dips at all times. The dips follow the morphological evolution of the flux ropes, in qualitative agreement with the observed filaments.

  2. Increased heat fluxes near a forest edge

    NARCIS (Netherlands)

    Klaassen, W; van Breugel, PB; Moors, EJ; Nieveen, JP

    2002-01-01

    Observations of sensible and latent heat flux above forest downwind of a forest edge show these fluxes to be larger than the available energy over the forest. The enhancement averages to 56 W m(-2), or 16% of the net radiation, at fetches less than 400 m, equivalent to fetch to height ratios less

  3. Increased heat fluxes near a forest edge

    NARCIS (Netherlands)

    Klaassen, W.; Breugel, van P.B.; Moors, E.J.; Nieveen, J.P.

    2002-01-01

    Observations of sensible and latent heat flux above forest downwind of a forest edge show these fluxes to be larger than the available energy over the forest. The enhancement averages to 56 W mm2, or 16 f the net radiation, at fetches less than 400 m, equivalent to fetch to height ratios less than

  4. Initiation of CMEs by Magnetic Flux Emergence

    Indian Academy of Sciences (India)

    The initiation of solar Coronal Mass Ejections (CMEs) is studied in the framework of numerical magnetohydrodynamics (MHD). The initial CME model includes a magnetic flux rope in spherical, axisymmetric geometry. The initial configuration consists of a magnetic flux rope embedded in a gravitationally stratified solar ...

  5. Crystal growth of emerald by flux method

    International Nuclear Information System (INIS)

    Inoue, Mikio; Narita, Eiichi; Okabe, Taijiro; Morishita, Toshihiko.

    1979-01-01

    Emerald crystals have been formed in two binary fluxes of Li 2 O-MoO 2 and Li 2 O-V 2 O 5 using the slow cooling method and the temperature gradient method under various conditions. In the flux of Li 2 O-MoO 3 carried out in the range of 2 -- 5 of molar ratios (MoO 3 /Li 2 O), emerald was crystallized in the temperature range from 750 to 950 0 C, and the suitable crystallization conditions were found to be the molar ratio of 3 -- 4 and the temperature about 900 0 C. In the flux of Li 2 O-V 2 O 5 carried out in the range of 1.7 -- 5 of molar ratios (V 2 O 5 /Li 2 O), emerald was crystallized in the temperature range from 900 to 1150 0 . The suitable crystals were obtained at the molar ratio of 3 and the temperature range of 1000 -- 1100 0 C. The crystallization temperature rised with an increase in the molar ratio of the both fluxes. The emeralds grown in two binary fluxes were transparent green, having the density of 2.68, the refractive index of 1.56, and the two distinct bands in the visible spectrum at 430 and 600nm. The emerald grown in Li 2 O-V 2 O 5 flux was more bluish green than that grown in Li 2 O-MoO 3 flux. The size of the spontaneously nucleated emerald grown in the former flux was larger than the latter, when crystallized by the slow cooling method. As for the solubility of beryl in the two fluxes, Li 2 O-V 2 O 5 flux was superior to Li 2 O-MoO 3 flux whose small solubility of SiO 2 caused an experimental problem to the temperature gradient method. The suitability of the two fluxes for the crystal growth of emerald by the flux method was discussed from the view point of various properties of above-mentioned two fluxes. (author)

  6. Flux Modulation in the Electrodynamic Loudspeaker

    DEFF Research Database (Denmark)

    Halvorsen, Morten; Tinggaard, Carsten; Agerkvist, Finn T.

    2015-01-01

    This paper discusses the effect of flux modulation in the electrodynamic loudspeaker with main focus on the effect on the force factor. A measurement setup to measure the AC flux modulation with static voice coil is explained and the measurements shows good consistency with FEA simulations....... Measurements of the generated AC flux modulation shows, that eddy currents are the main source to magnetic losses in form of phase lag and amplitude changes. Use of a copper cap shows a decrease in flux modulation amplitude at the expense of increased power losses. Finally, simulations show...... that there is a high dependency between the generated AC flux modulation from the voice coil and the AC force factor change....

  7. Plasma crowbars in cylindrical flux compression experiments

    International Nuclear Information System (INIS)

    Suter, L.J.

    1979-01-01

    We have done a series of one- and two-dimensional calculations of hard-core Z-pinch flux compression experiments in order to study the effect of a plasma on these systems. These calculations show that including a plasma can reduce the amount of flux lost during the compression. Flux losses to the outer wall of such experiments can be greatly reduced by a plasma conducting sheath which forms along the wall. This conducting sheath consists of a cold, dense high β, unmagnetized plasma which has enough pressure to balance a large field gradient. Flux which is lost into the center conductor is not effectively stopped by this plasma sheath until late in the implosion, at which time a layer similar to the one formed at the outer wall is created. Two-dimensionl simulations show that flux losses due to arching along the sliding contact of the experiment can be effectively stopped by the formation of a plasma conducting sheath

  8. Global observation-based diagnosis of soil moisture control on land surface flux partition

    Science.gov (United States)

    Gallego-Elvira, Belen; Taylor, Christopher M.; Harris, Phil P.; Ghent, Darren; Veal, Karen L.; Folwell, Sonja S.

    2016-04-01

    Soil moisture plays a central role in the partition of available energy at the land surface between sensible and latent heat flux to the atmosphere. As soils dry out, evapotranspiration becomes water-limited ("stressed"), and both land surface temperature (LST) and sensible heat flux rise as a result. This change in surface behaviour during dry spells directly affects critical processes in both the land and the atmosphere. Soil water deficits are often a precursor in heat waves, and they control where feedbacks on precipitation become significant. State-of-the-art global climate model (GCM) simulations for the Coupled Model Intercomparison Project Phase 5 (CMIP5) disagree on where and how strongly the surface energy budget is limited by soil moisture. Evaluation of GCM simulations at global scale is still a major challenge owing to the scarcity and uncertainty of observational datasets of land surface fluxes and soil moisture at the appropriate scale. Earth observation offers the potential to test how well GCM land schemes simulate hydrological controls on surface fluxes. In particular, satellite observations of LST provide indirect information about the surface energy partition at 1km resolution globally. Here, we present a potentially powerful methodology to evaluate soil moisture stress on surface fluxes within GCMs. Our diagnostic, Relative Warming Rate (RWR), is a measure of how rapidly the land warms relative to the overlying atmosphere during dry spells lasting at least 10 days. Under clear skies, this is a proxy for the change in sensible heat flux as soil dries out. We derived RWR from MODIS Terra and Aqua LST observations, meteorological re-analyses and satellite rainfall datasets. Globally we found that on average, the land warmed up during dry spells for 97% of the observed surface between 60S and 60N. For 73% of the area, the land warmed faster than the atmosphere (positive RWR), indicating water stressed conditions and increases in sensible heat flux

  9. OpenFLUX: efficient modelling software for 13C-based metabolic flux analysis

    Directory of Open Access Journals (Sweden)

    Nielsen Lars K

    2009-05-01

    Full Text Available Abstract Background The quantitative analysis of metabolic fluxes, i.e., in vivo activities of intracellular enzymes and pathways, provides key information on biological systems in systems biology and metabolic engineering. It is based on a comprehensive approach combining (i tracer cultivation on 13C substrates, (ii 13C labelling analysis by mass spectrometry and (iii mathematical modelling for experimental design, data processing, flux calculation and statistics. Whereas the cultivation and the analytical part is fairly advanced, a lack of appropriate modelling software solutions for all modelling aspects in flux studies is limiting the application of metabolic flux analysis. Results We have developed OpenFLUX as a user friendly, yet flexible software application for small and large scale 13C metabolic flux analysis. The application is based on the new Elementary Metabolite Unit (EMU framework, significantly enhancing computation speed for flux calculation. From simple notation of metabolic reaction networks defined in a spreadsheet, the OpenFLUX parser automatically generates MATLAB-readable metabolite and isotopomer balances, thus strongly facilitating model creation. The model can be used to perform experimental design, parameter estimation and sensitivity analysis either using the built-in gradient-based search or Monte Carlo algorithms or in user-defined algorithms. Exemplified for a microbial flux study with 71 reactions, 8 free flux parameters and mass isotopomer distribution of 10 metabolites, OpenFLUX allowed to automatically compile the EMU-based model from an Excel file containing metabolic reactions and carbon transfer mechanisms, showing it's user-friendliness. It reliably reproduced the published data and optimum flux distributions for the network under study were found quickly ( Conclusion We have developed a fast, accurate application to perform steady-state 13C metabolic flux analysis. OpenFLUX will strongly facilitate and

  10. Carbon Dioxide and Water Vapor Concentrations, Co-spectra and Fluxes from Latest Standardized Automated CO2/H2O Flux Systems versus Established Analyzer Models

    Science.gov (United States)

    Burba, G. G.; Kathilankal, J. C.; Begashaw, I.; Franzen, D.; Welles, J.; McDermitt, D. K.

    2017-12-01

    Spatial and temporal flux data coverage have improved significantly in recent years, due to standardization, automation and management of data collection, and better handling of the generated data. With more stations and networks, larger data streams from each station, and smaller operating budgets, modern tools are required to effectively and efficiently handle the entire process.These tools should produce standardized verifiable datasets, and provide a way to cross-share the standardized data with external collaborators to leverage available funding, and promote data analyses and publications. In 2015, new open-path and enclosed flux measurement systems1 were developed, based on established gas analyzer models2,3, with the goal of improving stability in the presence of contamination, refining temperature control and compensation, and providing more accurate gas concentration measurements. In 2017, the new open-path system was further refined to simplify hardware configuration, and to reduce power consumption and cost. Additionally, all new systems incorporate complete automated on-site flux calculations using EddyPro® Software4 run by a weatherized remotely-accessible microcomputer to provide standardized traceable data sets for fluxes and supporting variables. This presentation will describe details and results from the field tests of the new flux systems, in comparison to older models and reference instruments. References:1 Burba G., W. Miller, I. Begashaw, G. Fratini, F. Griessbaum, J. Kathilankal, L. Xu, D. Franz, E. Joseph, E. Larmanou, S. Miller, D. Papale, S. Sabbatini, T. Sachs, R. Sakai, D. McDermitt, 2017. Comparison of CO2 Concentrations, Co-spectra and Flux Measurements between Latest Standardized Automated CO2/H2O Flux Systems and Older Gas Analysers. 10th ICDC Conference, Switzerland: 21-25/08 2 Metzger, S., G. Burba, S. Burns, P. Blanken, J. Li, H. Luo, R. Zulueta, 2016. Optimization of an enclosed gas analyzer sampling system for measuring eddy

  11. Global High Resolution Sea Surface Flux Parameters From Multiple Satellites

    Science.gov (United States)

    Zhang, H.; Reynolds, R. W.; Shi, L.; Bates, J. J.

    2007-05-01

    Advances in understanding the coupled air-sea system and modeling of the ocean and atmosphere demand increasingly higher resolution data, such as air-sea fluxes of up to 3 hourly and every 50 km. These observational requirements can only be met by utilizing multiple satellite observations. Generation of such high resolution products from multiple-satellite and in-situ observations on an operational basis has been started at the U.S. National Oceanic and Atmospheric Administration (NOAA) National Climatic Data Center. Here we describe a few products that are directly related to the computation of turbulent air-sea fluxes. Sea surface wind speed has been observed from in-situ instruments and multiple satellites, with long-term observations ranging from one satellite in the mid 1987 to six or more satellites since mid 2002. A blended product with a global 0.25° grid and four snapshots per day has been produced for July 1987 to present, using a near Gaussian 3-D (x, y, t) interpolation to minimize aliases. Wind direction has been observed from fewer satellites, thus for the blended high resolution vector winds and wind stresses, the directions are taken from the NCEP Re-analysis 2 (operationally run near real time) for climate consistency. The widely used Reynolds Optimum Interpolation SST analysis has been improved with higher resolutions (daily and 0.25°). The improvements use both infrared and microwave satellite data that are bias-corrected by in- situ observations for the period 1985 to present. The new versions provide very significant improvements in terms of resolving ocean features such as the meandering of the Gulf Stream, the Aghulas Current, the equatorial jets and other fronts. The Ta and Qa retrievals are based on measurements from the AMSU sounder onboard the NOAA satellites. Ta retrieval uses AMSU-A data, while Qa retrieval uses both AMSU-A and AMSU-B observations. The retrieval algorithms are developed using the neural network approach. Training

  12. A novel dataset for real-life evaluation of facial expression recognition methodologies

    NARCIS (Netherlands)

    Siddiqi, Muhammad Hameed; Ali, Maqbool; Idris, Muhammad; Banos Legran, Oresti; Lee, Sungyoung; Choo, Hyunseung

    2016-01-01

    One limitation seen among most of the previous methods is that they were evaluated under settings that are far from real-life scenarios. The reason is that the existing facial expression recognition (FER) datasets are mostly pose-based and assume a predefined setup. The expressions in these datasets

  13. Full-Scale Approximations of Spatio-Temporal Covariance Models for Large Datasets

    KAUST Repository

    Zhang, Bohai; Sang, Huiyan; Huang, Jianhua Z.

    2014-01-01

    of dataset and application of such models is not feasible for large datasets. This article extends the full-scale approximation (FSA) approach by Sang and Huang (2012) to the spatio-temporal context to reduce computational complexity. A reversible jump Markov

  14. Gridded precipitation dataset for the Rhine basin made with the genRE interpolation method

    NARCIS (Netherlands)

    Osnabrugge, van B.; Uijlenhoet, R.

    2017-01-01

    A high resolution (1.2x1.2km) gridded precipitation dataset with hourly time step that covers the whole Rhine basin for the period 1997-2015. Made from gauge data with the genRE interpolation scheme. See "genRE: A method to extend gridded precipitation climatology datasets in near real-time for

  15. TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking in the Wild

    KAUST Repository

    Mü ller, Matthias; Bibi, Adel Aamer; Giancola, Silvio; Al-Subaihi, Salman; Ghanem, Bernard

    2018-01-01

    Despite the numerous developments in object tracking, further development of current tracking algorithms is limited by small and mostly saturated datasets. As a matter of fact, data-hungry trackers based on deep-learning currently rely on object detection datasets due to the scarcity of dedicated large-scale tracking datasets. In this work, we present TrackingNet, the first large-scale dataset and benchmark for object tracking in the wild. We provide more than 30K videos with more than 14 million dense bounding box annotations. Our dataset covers a wide selection of object classes in broad and diverse context. By releasing such a large-scale dataset, we expect deep trackers to further improve and generalize. In addition, we introduce a new benchmark composed of 500 novel videos, modeled with a distribution similar to our training dataset. By sequestering the annotation of the test set and providing an online evaluation server, we provide a fair benchmark for future development of object trackers. Deep trackers fine-tuned on a fraction of our dataset improve their performance by up to 1.6% on OTB100 and up to 1.7% on TrackingNet Test. We provide an extensive benchmark on TrackingNet by evaluating more than 20 trackers. Our results suggest that object tracking in the wild is far from being solved.

  16. Omicseq: a web-based search engine for exploring omics datasets

    Science.gov (United States)

    Sun, Xiaobo; Pittard, William S.; Xu, Tianlei; Chen, Li; Zwick, Michael E.; Jiang, Xiaoqian; Wang, Fusheng

    2017-01-01

    Abstract The development and application of high-throughput genomics technologies has resulted in massive quantities of diverse omics data that continue to accumulate rapidly. These rich datasets offer unprecedented and exciting opportunities to address long standing questions in biomedical research. However, our ability to explore and query the content of diverse omics data is very limited. Existing dataset search tools rely almost exclusively on the metadata. A text-based query for gene name(s) does not work well on datasets wherein the vast majority of their content is numeric. To overcome this barrier, we have developed Omicseq, a novel web-based platform that facilitates the easy interrogation of omics datasets holistically to improve ‘findability’ of relevant data. The core component of Omicseq is trackRank, a novel algorithm for ranking omics datasets that fully uses the numerical content of the dataset to determine relevance to the query entity. The Omicseq system is supported by a scalable and elastic, NoSQL database that hosts a large collection of processed omics datasets. In the front end, a simple, web-based interface allows users to enter queries and instantly receive search results as a list of ranked datasets deemed to be the most relevant. Omicseq is freely available at http://www.omicseq.org. PMID:28402462

  17. Document Questionnaires and Datasets with DDI: A Hands-On Introduction with Colectica

    OpenAIRE

    Iverson, Jeremy; Smith, Dan

    2018-01-01

    This workshop offers a hands-on, practical approach to creating and documenting both surveys and datasets with DDI and Colectica. Participants will build and field a DDI-driven survey using their own questions or samples provided in the workshop. They will then ingest, annotate, and publish DDI dataset descriptions using the collected survey data.

  18. An integrated pan-tropical biomass map using multiple reference datasets

    NARCIS (Netherlands)

    Avitabile, V.; Herold, M.; Heuvelink, G.B.M.; Lewis, S.L.; Phillips, O.L.; Asner, G.P.; Armston, J.; Asthon, P.; Banin, L.F.; Bayol, N.; Berry, N.; Boeckx, P.; Jong, De B.; Devries, B.; Girardin, C.; Kearsley, E.; Lindsell, J.A.; Lopez-gonzalez, G.; Lucas, R.; Malhi, Y.; Morel, A.; Mitchard, E.; Nagy, L.; Qie, L.; Quinones, M.; Ryan, C.M.; Slik, F.; Sunderland, T.; Vaglio Laurin, G.; Valentini, R.; Verbeeck, H.; Wijaya, A.; Willcock, S.

    2016-01-01

    We combined two existing datasets of vegetation aboveground biomass (AGB) (Proceedings of the National Academy of Sciences of the United States of America, 108, 2011, 9899; Nature Climate Change, 2, 2012, 182) into a pan-tropical AGB map at 1-km resolution using an independent reference dataset of

  19. Large-scale Labeled Datasets to Fuel Earth Science Deep Learning Applications

    Science.gov (United States)

    Maskey, M.; Ramachandran, R.; Miller, J.

    2017-12-01

    Deep learning has revolutionized computer vision and natural language processing with various algorithms scaled using high-performance computing. However, generic large-scale labeled datasets such as the ImageNet are the fuel that drives the impressive accuracy of deep learning results. Large-scale labeled datasets already exist in domains such as medical science, but creating them in the Earth science domain is a challenge. While there are ways to apply deep learning using limited labeled datasets, there is a need in the Earth sciences for creating large-scale labeled datasets for benchmarking and scaling deep learning applications. At the NASA Marshall Space Flight Center, we are using deep learning for a variety of Earth science applications where we have encountered the need for large-scale labeled datasets. We will discuss our approaches for creating such datasets and why these datasets are just as valuable as deep learning algorithms. We will also describe successful usage of these large-scale labeled datasets with our deep learning based applications.

  20. SAR image classification based on CNN in real and simulation datasets

    Science.gov (United States)

    Peng, Lijiang; Liu, Ming; Liu, Xiaohua; Dong, Liquan; Hui, Mei; Zhao, Yuejin

    2018-04-01

    Convolution neural network (CNN) has made great success in image classification tasks. Even in the field of synthetic aperture radar automatic target recognition (SAR-ATR), state-of-art results has been obtained by learning deep representation of features on the MSTAR benchmark. However, the raw data of MSTAR have shortcomings in training a SAR-ATR model because of high similarity in background among the SAR images of each kind. This indicates that the CNN would learn the hierarchies of features of backgrounds as well as the targets. To validate the influence of the background, some other SAR images datasets have been made which contains the simulation SAR images of 10 manufactured targets such as tank and fighter aircraft, and the backgrounds of simulation SAR images are sampled from the whole original MSTAR data. The simulation datasets contain the dataset that the backgrounds of each kind images correspond to the one kind of backgrounds of MSTAR targets or clutters and the dataset that each image shares the random background of whole MSTAR targets or clutters. In addition, mixed datasets of MSTAR and simulation datasets had been made to use in the experiments. The CNN architecture proposed in this paper are trained on all datasets mentioned above. The experimental results shows that the architecture can get high performances on all datasets even the backgrounds of the images are miscellaneous, which indicates the architecture can learn a good representation of the targets even though the drastic changes on background.