WorldWideScience

Sample records for standards-compatible geospatial catalogue

  1. Grid Enabled Geospatial Catalogue Web Service

    Science.gov (United States)

    Chen, Ai-Jun; Di, Li-Ping; Wei, Ya-Xing; Liu, Yang; Bui, Yu-Qi; Hu, Chau-Min; Mehrotra, Piyush

    2004-01-01

    Geospatial Catalogue Web Service is a vital service for sharing and interoperating volumes of distributed heterogeneous geospatial resources, such as data, services, applications, and their replicas over the web. Based on the Grid technology and the Open Geospatial Consortium (0GC) s Catalogue Service - Web Information Model, this paper proposes a new information model for Geospatial Catalogue Web Service, named as GCWS which can securely provides Grid-based publishing, managing and querying geospatial data and services, and the transparent access to the replica data and related services under the Grid environment. This information model integrates the information model of the Grid Replica Location Service (RLS)/Monitoring & Discovery Service (MDS) with the information model of OGC Catalogue Service (CSW), and refers to the geospatial data metadata standards from IS0 19115, FGDC and NASA EOS Core System and service metadata standards from IS0 191 19 to extend itself for expressing geospatial resources. Using GCWS, any valid geospatial user, who belongs to an authorized Virtual Organization (VO), can securely publish and manage geospatial resources, especially query on-demand data in the virtual community and get back it through the data-related services which provide functions such as subsetting, reformatting, reprojection etc. This work facilitates the geospatial resources sharing and interoperating under the Grid environment, and implements geospatial resources Grid enabled and Grid technologies geospatial enabled. It 2!so makes researcher to focus on science, 2nd not cn issues with computing ability, data locztic, processir,g and management. GCWS also is a key component for workflow-based virtual geospatial data producing.

  2. Restful Implementation of Catalogue Service for Geospatial Data Provenance

    Science.gov (United States)

    Jiang, L. C.; Yue, P.; Lu, X. C.

    2013-10-01

    Provenance, also known as lineage, is important in understanding the derivation history of data products. Geospatial data provenance helps data consumers to evaluate the quality and reliability of geospatial data. In a service-oriented environment, where data are often consumed or produced by distributed services, provenance could be managed by following the same service-oriented paradigm. The Open Geospatial Consortium (OGC) Catalogue Service for the Web (CSW) is used for the registration and query of geospatial data provenance by extending ebXML Registry Information Model (ebRIM). Recent advance of the REpresentational State Transfer (REST) paradigm has shown great promise for the easy integration of distributed resources. RESTful Web Service aims to provide a standard way for Web clients to communicate with servers based on REST principles. The existing approach for provenance catalogue service could be improved by adopting the RESTful design. This paper presents the design and implementation of a catalogue service for geospatial data provenance following RESTful architecture style. A middleware named REST Converter is added on the top of the legacy catalogue service to support a RESTful style interface. The REST Converter is composed of a resource request dispatcher and six resource handlers. A prototype service is developed to demonstrate the applicability of the approach.

  3. Distributed Multi-interface Catalogue for Geospatial Data

    Science.gov (United States)

    Nativi, S.; Bigagli, L.; Mazzetti, P.; Mattia, U.; Boldrini, E.

    2007-12-01

    Several geosciences communities (e.g. atmospheric science, oceanography, hydrology) have developed tailored data and metadata models and service protocol specifications for enabling online data discovery, inventory, evaluation, access and download. These specifications are conceived either profiling geospatial information standards or extending the well-accepted geosciences data models and protocols in order to capture more semantics. These artifacts have generated a set of related catalog -and inventory services- characterizing different communities, initiatives and projects. In fact, these geospatial data catalogs are discovery and access systems that use metadata as the target for query on geospatial information. The indexed and searchable metadata provide a disciplined vocabulary against which intelligent geospatial search can be performed within or among communities. There exists a clear need to conceive and achieve solutions to implement interoperability among geosciences communities, in the context of the more general geospatial information interoperability framework. Such solutions should provide search and access capabilities across catalogs, inventory lists and their registered resources. Thus, the development of catalog clearinghouse solutions is a near-term challenge in support of fully functional and useful infrastructures for spatial data (e.g. INSPIRE, GMES, NSDI, GEOSS). This implies the implementation of components for query distribution and virtual resource aggregation. These solutions must implement distributed discovery functionalities in an heterogeneous environment, requiring metadata profiles harmonization as well as protocol adaptation and mediation. We present a catalog clearinghouse solution for the interoperability of several well-known cataloguing systems (e.g. OGC CSW, THREDDS catalog and data services). The solution implements consistent resource discovery and evaluation over a dynamic federation of several well-known cataloguing and

  4. The Future of Geospatial Standards

    Science.gov (United States)

    Bermudez, L. E.; Simonis, I.

    2016-12-01

    The OGC is an international not-for-profit standards development organization (SDO) committed to making quality standards for the geospatial community. A community of more than 500 member organizations with more than 6,000 people registered at the OGC communication platform drives the development of standards that are freely available for anyone to use and to improve sharing of the world's geospatial data. OGC standards are applied in a variety of application domains including Environment, Defense and Intelligence, Smart Cities, Aviation, Disaster Management, Agriculture, Business Development and Decision Support, and Meteorology. Profiles help to apply information models to different communities, thus adapting to particular needs of that community while ensuring interoperability by using common base models and appropriate support services. Other standards address orthogonal aspects such as handling of Big Data, Crowd-sourced information, Geosemantics, or container for offline data usage. Like most SDOs, the OGC develops and maintains standards through a formal consensus process under the OGC Standards Program (OGC-SP) wherein requirements and use cases are discussed in forums generally open to the public (Domain Working Groups, or DWGs), and Standards Working Groups (SWGs) are established to create standards. However, OGC is unique among SDOs in that it also operates the OGC Interoperability Program (OGC-IP) to provide real-world testing of existing and proposed standards. The OGC-IP is considered the experimental playground, where new technologies are researched and developed in a user-driven process. Its goal is to prototype, test, demonstrate, and promote OGC Standards in a structured environment. Results from the OGC-IP often become requirements for new OGC standards or identify deficiencies in existing OGC standards that can be addressed. This presentation will provide an analysis of the work advanced in the OGC consortium including standards and testbeds

  5. The new OGC Catalogue Services 3.0 specification - status of work

    Science.gov (United States)

    Bigagli, Lorenzo; Voges, Uwe

    2013-04-01

    We report on the work of the Open Geospatial Consortium Catalogue Services 3.0 Standards Working Group (OGC Cat 3.0 SWG for short), started in March 2008, with the purpose to process change requests on the Catalogue Services 2.0.2 Implementation Specification (OGC 07-006r1) and produce a revised version thereof, comprising the related XML schemas and abstract test suite. The work was initially intended as a minor revision (version 2.1), but later retargeted as a major update of the standard and rescheduled (the anticipated roadmap ended in 2008). The target audience of Catalogue Services 3.0 includes: • Implementors of catalogue services solutions. • Designers and developers of catalogue services profiles. • Providers/users of catalogue services. The two main general areas of enhancement included: restructuring the specification document according to the OGC standard for modular specifications (OGC 08-131r3, also known as Core and Extension model); incorporating the current mass-market technologies for discovery on the Web, namely OpenSearch. The document was initially split into four parts: the general model and the three protocol bindings HTTP, Z39.50, and CORBA. The CORBA binding, which was very rarely implemented, and the Z39.50 binding have later been dropped. Parts of the Z39.50 binding, namely Search/Retrieve via URL (SRU; same semantics as Z39.50, but stateless), have been provided as a discussion paper (OGC 12-082) for possibly developing a future SRU profile. The Catalogue Services 3.0 specification is structured as follows: • Part 1: General Model (Core) • Part 2: HTTP Protocol Binding (CSW) In CSW, the GET/KVP encoding is mandatory. The POST/XML encoding is optional. SOAP is supported as a special case of the POST/XML encoding. OpenSearch must always be supported, regardless of the implemented profiles, along with the OpenSearch Geospatial and Temporal Extensions (OGC 10-032r2). The latter specifies spatial (e.g. point-plus-radius, bounding

  6. English translations of German standards. Catalogue 1988. 24. ed.

    International Nuclear Information System (INIS)

    1988-01-01

    The catalogue contains a list of all currently available English translations of DIN standards, and of English translations of DIN handbooks, a numerical index, an alphabetical index, and an index of DIN EN, DIN IEC, DIN ISO standards, LN and VG standards. Some useful information on standards work in Germany and on the activities of DIN Deutsches Institut fuer Normung e.V. is given. (orig./HP)

  7. GeoSearch: A lightweight broking middleware for geospatial resources discovery

    Science.gov (United States)

    Gui, Z.; Yang, C.; Liu, K.; Xia, J.

    2012-12-01

    With petabytes of geodata, thousands of geospatial web services available over the Internet, it is critical to support geoscience research and applications by finding the best-fit geospatial resources from the massive and heterogeneous resources. Past decades' developments witnessed the operation of many service components to facilitate geospatial resource management and discovery. However, efficient and accurate geospatial resource discovery is still a big challenge due to the following reasons: 1)The entry barriers (also called "learning curves") hinder the usability of discovery services to end users. Different portals and catalogues always adopt various access protocols, metadata formats and GUI styles to organize, present and publish metadata. It is hard for end users to learn all these technical details and differences. 2)The cost for federating heterogeneous services is high. To provide sufficient resources and facilitate data discovery, many registries adopt periodic harvesting mechanism to retrieve metadata from other federated catalogues. These time-consuming processes lead to network and storage burdens, data redundancy, and also the overhead of maintaining data consistency. 3)The heterogeneous semantics issues in data discovery. Since the keyword matching is still the primary search method in many operational discovery services, the search accuracy (precision and recall) is hard to guarantee. Semantic technologies (such as semantic reasoning and similarity evaluation) offer a solution to solve these issues. However, integrating semantic technologies with existing service is challenging due to the expandability limitations on the service frameworks and metadata templates. 4)The capabilities to help users make final selection are inadequate. Most of the existing search portals lack intuitive and diverse information visualization methods and functions (sort, filter) to present, explore and analyze search results. Furthermore, the presentation of the value

  8. Energetic materials standards – Chemical compatibility

    NARCIS (Netherlands)

    Tuukkanen, I.M.; Bouma, R.H.B.

    2014-01-01

    Subgroup A Energetic Materials Team, SG/A (EMT), develops and maintains standards that are relevant to all life-cycle phases of ammunition/weapon systems. STANAG 4147 is the standard regarding chemical compatibility of explosives with munition components, and is a document of prime importance.

  9. Bridging IMO e-Navigation Policy and Offshore Oil and Gas Operations through Geospatial Standards

    Directory of Open Access Journals (Sweden)

    Filipe Modesto Da Rocha

    2016-04-01

    Full Text Available In offshore industry activities, the suitable onboard provision of assets location and geospatial marine information during operations is essential. Currently, most companies use its own data structures, resulting in incompatibility between processes. In order to promote the data exchange, oil and gas industry associations have pursued initiatives to standardize spatial information. In turn, the IMO - International Maritime Organization - started the implementation of e-Navigation policy, which is the standardization of technologies and protocols applied to maritime information and navigation. This paper shows relationship and integration points between maritime activities of oil and gas industry and e-Navigation technologies and processes, highlighting geospatial information. This paper also preludes out an initiative for a suitable product specification for the offshore oil and gas industry, compliant with e-Navigation and IHO S-100 international standards.

  10. Catalogue of EDF`s technical specifications (HN reference). Centre of standardization; Catalogue des specifications techniques EDF (reference HN). Centre de normalisation

    Energy Technology Data Exchange (ETDEWEB)

    1998-12-31

    This document edited by Electricite de France (EdF), is a catalogue of the French standard documents relative to any type of electrical material and equipment and which contain the technical specifications of these materials and equipments. A brief description of these specifications is given for each type of material or equipment listed. (J.S.)

  11. On the moroccan tsunami catalogue

    Directory of Open Access Journals (Sweden)

    F. Kaabouben

    2009-07-01

    Full Text Available A primary tool for regional tsunami hazard assessment is a reliable historical and instrumental catalogue of events. Morocco by its geographical situation, with two marine sides, stretching along the Atlantic coast to the west and along the Mediterranean coast to the north, is the country of Western Africa most exposed to the risk of tsunamis. Previous information on tsunami events affecting Morocco are included in the Iberian and/or the Mediterranean lists of tsunami events, as it is the case of the European GITEC Tsunami Catalogue, but there is a need to organize this information in a dataset and to assess the likelihood of claimed historical tsunamis in Morocco. Due to the fact that Moroccan sources are scarce, this compilation rely on historical documentation from neighbouring countries (Portugal and Spain and so the compatibility between the new tsunami catalogue presented here and those that correspond to the same source areas is also discussed.

  12. Interpreting the ASTM 'content standard for digital geospatial metadata'

    Science.gov (United States)

    Nebert, Douglas D.

    1996-01-01

    ASTM and the Federal Geographic Data Committee have developed a content standard for spatial metadata to facilitate documentation, discovery, and retrieval of digital spatial data using vendor-independent terminology. Spatial metadata elements are identifiable quality and content characteristics of a data set that can be tied to a geographic location or area. Several Office of Management and Budget Circulars and initiatives have been issued that specify improved cataloguing of and accessibility to federal data holdings. An Executive Order further requires the use of the metadata content standard to document digital spatial data sets. Collection and reporting of spatial metadata for field investigations performed for the federal government is an anticipated requirement. This paper provides an overview of the draft spatial metadata content standard and a description of how the standard could be applied to investigations collecting spatially-referenced field data.

  13. Alcohol promotions in Australian supermarket catalogues.

    Science.gov (United States)

    Johnston, Robyn; Stafford, Julia; Pierce, Hannah; Daube, Mike

    2017-07-01

    In Australia, most alcohol is sold as packaged liquor from off-premises retailers, a market increasingly dominated by supermarket chains. Competition between retailers may encourage marketing approaches, for example, discounting, that evidence indicates contribute to alcohol-related harms. This research documented the nature and variety of promotional methods used by two major supermarket retailers to promote alcohol products in their supermarket catalogues. Weekly catalogues from the two largest Australian supermarket chains were reviewed for alcohol-related content over 12 months. Alcohol promotions were assessed for promotion type, product type, number of standard drinks, purchase price and price/standard drink. Each store catalogue included, on average, 13 alcohol promotions/week, with price-based promotions most common. Forty-five percent of promotions required the purchase of multiple alcohol items. Wine was the most frequently promoted product (44%), followed by beer (24%) and spirits (18%). Most (99%) wine cask (2-5 L container) promotions required multiple (two to three) casks to be purchased. The average number of standard drinks required to be purchased to participate in catalogue promotions was 31.7 (SD = 24.9; median = 23.1). The median price per standard drink was $1.49 (range $0.19-$9.81). Cask wines had the lowest cost per standard drink across all product types. Supermarket catalogues' emphasis on low prices/high volumes of alcohol reflects that retailers are taking advantage of limited restrictions on off-premise sales and promotion, which allow them to approach market competition in ways that may increase alcohol-related harms in consumers. Regulation of alcohol marketing should address retailer catalogue promotions. [Johnston R, Stafford J, Pierce H, Daube M. Alcohol promotions in Australian supermarket catalogues. Drug Alcohol Rev 2017;36:456-463]. © 2016 Australasian Professional Society on Alcohol and other Drugs.

  14. Cloud Computing for Geosciences--GeoCloud for standardized geospatial service platforms (Invited)

    Science.gov (United States)

    Nebert, D. D.; Huang, Q.; Yang, C.

    2013-12-01

    The 21st century geoscience faces challenges of Big Data, spike computing requirements (e.g., when natural disaster happens), and sharing resources through cyberinfrastructure across different organizations (Yang et al., 2011). With flexibility and cost-efficiency of computing resources a primary concern, cloud computing emerges as a promising solution to provide core capabilities to address these challenges. Many governmental and federal agencies are adopting cloud technologies to cut costs and to make federal IT operations more efficient (Huang et al., 2010). However, it is still difficult for geoscientists to take advantage of the benefits of cloud computing to facilitate the scientific research and discoveries. This presentation reports using GeoCloud to illustrate the process and strategies used in building a common platform for geoscience communities to enable the sharing, integration of geospatial data, information and knowledge across different domains. GeoCloud is an annual incubator project coordinated by the Federal Geographic Data Committee (FGDC) in collaboration with the U.S. General Services Administration (GSA) and the Department of Health and Human Services. It is designed as a staging environment to test and document the deployment of a common GeoCloud community platform that can be implemented by multiple agencies. With these standardized virtual geospatial servers, a variety of government geospatial applications can be quickly migrated to the cloud. In order to achieve this objective, multiple projects are nominated each year by federal agencies as existing public-facing geospatial data services. From the initial candidate projects, a set of common operating system and software requirements was identified as the baseline for platform as a service (PaaS) packages. Based on these developed common platform packages, each project deploys and monitors its web application, develops best practices, and documents cost and performance information. This

  15. Geoportale del Consorzio LaMMA Disseminazione di dati meteo in near real-time tramite standard OGC e software Open Source

    Directory of Open Access Journals (Sweden)

    Simone Giannechini

    2014-02-01

    Full Text Available This paper describes the spatial data infrastructure (SDI used by the LaMMA Consortium - Environmental Mod elling and Monitoring Laboratory for Sustainable Developm ent of Tuscany Region for sharing, viewing and cataloguing (metadata and related information all geospatial data that are daily proc essed and used op erationally in many meteorological and environmental app lications.The SDI was develop ed using Open Source technologies, mo reover the geospatial data has been imp lemented through protoco ls based on ogc (Open Geospatial Consortium standards such as WMS, WFS and CSW. Geoserver was used for disseminating geospatial data and maps through OGC WMS and WFS protoco ls while GeoNetwork was used as the cataloguing and search po rtal through also the CSW protocol; eventually MapStore was used to implement the mash-up front-end.The innovative aspect of this po rtal is the fact that it currently is ingesting, fusing and disseminating geospatial data related to the MetOcfield from various sources in near real-time in a comp rehensive manner that allows users to create add ed value visualizations for the support of operational use cases as well as to access and download underlying data (where app licable.

  16. Operational Marine Data Acquisition and Delivery Powered by Web and Geospatial Standards

    Science.gov (United States)

    Thomas, R.; Buck, J. J. H.

    2015-12-01

    As novel sensor types and new platforms are deployed to monitor the global oceans, the volumes of scientific and environmental data collected in the marine context are rapidly growing. In order to use these data in both the traditional operational modes and in innovative "Big Data" applications the data must be readily understood by software agents. One approach to achieving this is the application of both World Wide Web and Open Geospatial Consortium standards: namely Linked Data1 and Sensor Web Enablement2 (SWE). The British Oceanographic Data Centre (BODC) is adopting this strategy in a number of European Commission funded projects (NETMAR; SenseOCEAN; Ocean Data Interoperability Platform - ODIP; and AtlantOS) to combine its existing data archiving architecture with SWE components (such as Sensor Observation Services) and a Linked Data interface. These will evolve the data management and data transfer from a process that requires significant manual intervention to an automated operational process enabling the rapid, standards-based, ingestion and delivery of data. This poster will show the current capabilities of BODC and the status of on-going implementation of this strategy. References1. World Wide Web Consortium. (2013). Linked Data. Available:http://www.w3.org/standards/semanticweb/data. Last accessed 7th April 20152. Open Geospatial Consortium. (2014). Sensor Web Enablement (SWE). Available:http://www.opengeospatial.org/ogc/markets-technologies/swe. Last accessed 8th October 2014

  17. Practical cataloguing AACR, RDA and MARC 21

    CERN Document Server

    Welsh, Anne

    2012-01-01

    Written at a time of transition in international cataloguing, this book provides cataloguers and students with a background in general cataloguing principles, the code (AACR2) and format (MARC 21) and the new standard (RDA). It provides library managers with an overview of the development of RDA in order to equip them to make the transition.

  18. SemantGeo: Powering Ecological and Environment Data Discovery and Search with Standards-Based Geospatial Reasoning

    Science.gov (United States)

    Seyed, P.; Ashby, B.; Khan, I.; Patton, E. W.; McGuinness, D. L.

    2013-12-01

    Recent efforts to create and leverage standards for geospatial data specification and inference include the GeoSPARQL standard, Geospatial OWL ontologies (e.g., GAZ, Geonames), and RDF triple stores that support GeoSPARQL (e.g., AllegroGraph, Parliament) that use RDF instance data for geospatial features of interest. However, there remains a gap on how best to fuse software engineering best practices and GeoSPARQL within semantic web applications to enable flexible search driven by geospatial reasoning. In this abstract we introduce the SemantGeo module for the SemantEco framework that helps fill this gap, enabling scientists find data using geospatial semantics and reasoning. SemantGeo provides multiple types of geospatial reasoning for SemantEco modules. The server side implementation uses the Parliament SPARQL Endpoint accessed via a Tomcat servlet. SemantGeo uses the Google Maps API for user-specified polygon construction and JsTree for providing containment and categorical hierarchies for search. SemantGeo uses GeoSPARQL for spatial reasoning alone and in concert with RDFS/OWL reasoning capabilities to determine, e.g., what geofeatures are within, partially overlap with, or within a certain distance from, a given polygon. We also leverage qualitative relationships defined by the Gazetteer ontology that are composites of spatial relationships as well as administrative designations or geophysical phenomena. We provide multiple mechanisms for exploring data, such as polygon (map-based) and named-feature (hierarchy-based) selection, that enable flexible search constraints using boolean combination of selections. JsTree-based hierarchical search facets present named features and include a 'part of' hierarchy (e.g., measurement-site-01, Lake George, Adirondack Region, NY State) and type hierarchies (e.g., nodes in the hierarchy for WaterBody, Park, MeasurementSite), depending on the ';axis of choice' option selected. Using GeoSPARQL and aforementioned ontology

  19. Catalogue and classification of technical safety standards, rules and regulations for nuclear power reactors and nuclear fuel cycle facilities

    International Nuclear Information System (INIS)

    Fichtner, N.; Becker, K.; Bashir, M.

    1977-01-01

    The present report is an up-dated version of the report 'Catalogue and Classification of Technical Safety Rules for Light-water Reactors and Reprocessing Plants' edited under code No EUR 5362e, August 1975. Like the first version of the report, it constitutes a catalogue and classification of standards, rules and regulations on land-based nuclear power reactors and fuel cycle facilities. The reasons for the classification system used are given and discussed

  20. Interoperability in planetary research for geospatial data analysis

    Science.gov (United States)

    Hare, Trent M.; Rossi, Angelo P.; Frigeri, Alessandro; Marmo, Chiara

    2018-01-01

    For more than a decade there has been a push in the planetary science community to support interoperable methods for accessing and working with geospatial data. Common geospatial data products for planetary research include image mosaics, digital elevation or terrain models, geologic maps, geographic location databases (e.g., craters, volcanoes) or any data that can be tied to the surface of a planetary body (including moons, comets or asteroids). Several U.S. and international cartographic research institutions have converged on mapping standards that embrace standardized geospatial image formats, geologic mapping conventions, U.S. Federal Geographic Data Committee (FGDC) cartographic and metadata standards, and notably on-line mapping services as defined by the Open Geospatial Consortium (OGC). The latter includes defined standards such as the OGC Web Mapping Services (simple image maps), Web Map Tile Services (cached image tiles), Web Feature Services (feature streaming), Web Coverage Services (rich scientific data streaming), and Catalog Services for the Web (data searching and discoverability). While these standards were developed for application to Earth-based data, they can be just as valuable for planetary domain. Another initiative, called VESPA (Virtual European Solar and Planetary Access), will marry several of the above geoscience standards and astronomy-based standards as defined by International Virtual Observatory Alliance (IVOA). This work outlines the current state of interoperability initiatives in use or in the process of being researched within the planetary geospatial community.

  1. A new version of the European tsunami catalogue: updating and revision

    Directory of Open Access Journals (Sweden)

    S. Tinti

    2001-01-01

    Full Text Available A new version of the European catalogue of tsunamis is presented here. It differs from the latest release of the catalogue that was produced in 1998 and is known as GITEC tsunami catalogue in some important aspects. In the first place, it is a database built on the Visual FoxPro 6.0 DBMS that can be used and maintained under the PC operating systems currently available. Conversely, the GITEC catalogue was compatible only with Windows 95 and older PC platforms. In the second place, it is enriched by new facilities and a new type of data, such as a database of pictures that can be accessed easily from the main screen of the catalogue. Thirdly, it has been updated by including the newly published references. Minute and painstaking search for new data has been undertaken to re-evaluate cases that were not included in the GITEC catalogue, though they were mentioned in previous catalogues; the exclusion was motivated by a lack of data. This last work has focused so far on Italian cases of the last two centuries. The result is that at least two events have been found which deserve inclusion in the new catalogue: one occurred in 1809 in the Gulf of La Spezia, and the other occurred in 1940 in the Gulf of Palermo. Two further events are presently under investigation.

  2. An Effective Framework for Distributed Geospatial Query Processing in Grids

    Directory of Open Access Journals (Sweden)

    CHEN, B.

    2010-08-01

    Full Text Available The emergence of Internet has greatly revolutionized the way that geospatial information is collected, managed, processed and integrated. There are several important research issues to be addressed for distributed geospatial applications. First, the performance of geospatial applications is needed to be considered in the Internet environment. In this regard, the Grid as an effective distributed computing paradigm is a good choice. The Grid uses a series of middleware to interconnect and merge various distributed resources into a super-computer with capability of high performance computation. Secondly, it is necessary to ensure the secure use of independent geospatial applications in the Internet environment. The Grid just provides the utility of secure access to distributed geospatial resources. Additionally, it makes good sense to overcome the heterogeneity between individual geospatial information systems in Internet. The Open Geospatial Consortium (OGC proposes a number of generalized geospatial standards e.g. OGC Web Services (OWS to achieve interoperable access to geospatial applications. The OWS solution is feasible and widely adopted by both the academic community and the industry community. Therefore, we propose an integrated framework by incorporating OWS standards into Grids. Upon the framework distributed geospatial queries can be performed in an interoperable, high-performance and secure Grid environment.

  3. Cloud computing geospatial application for water resources based on free and open source software and open standards - a prototype

    Science.gov (United States)

    Delipetrev, Blagoj

    2016-04-01

    Presently, most of the existing software is desktop-based, designed to work on a single computer, which represents a major limitation in many ways, starting from limited computer processing, storage power, accessibility, availability, etc. The only feasible solution lies in the web and cloud. This abstract presents research and development of a cloud computing geospatial application for water resources based on free and open source software and open standards using hybrid deployment model of public - private cloud, running on two separate virtual machines (VMs). The first one (VM1) is running on Amazon web services (AWS) and the second one (VM2) is running on a Xen cloud platform. The presented cloud application is developed using free and open source software, open standards and prototype code. The cloud application presents a framework how to develop specialized cloud geospatial application that needs only a web browser to be used. This cloud application is the ultimate collaboration geospatial platform because multiple users across the globe with internet connection and browser can jointly model geospatial objects, enter attribute data and information, execute algorithms, and visualize results. The presented cloud application is: available all the time, accessible from everywhere, it is scalable, works in a distributed computer environment, it creates a real-time multiuser collaboration platform, the programing languages code and components are interoperable, and it is flexible in including additional components. The cloud geospatial application is implemented as a specialized water resources application with three web services for 1) data infrastructure (DI), 2) support for water resources modelling (WRM), 3) user management. The web services are running on two VMs that are communicating over the internet providing services to users. The application was tested on the Zletovica river basin case study with concurrent multiple users. The application is a state

  4. The Euro-Mediterranean Tsunami Catalogue

    Directory of Open Access Journals (Sweden)

    Alessandra Maramai

    2014-08-01

    Full Text Available A unified catalogue containing 290 tsunamis generated in the European and Mediterranean seas since 6150 B.C. to current days is presented. It is the result of a systematic and detailed review of all the regional catalogues available in literature covering the study area, each of them having their own format and level of accuracy. The realization of a single catalogue covering a so wide area and involving several countries was a complex task that posed a series of challenges, being the standardization and the quality of the data the most demanding. A “reliability” value was used to rate equally the quality of the data for each event and this parameter was assigned based on the trustworthiness of the information related to the generating cause, the tsunami description accuracy and also on the availability of coeval bibliographical sources. Following these criteria we included in the catalogue events whose reliability ranges from 0 (“very improbable tsunami” to 4 (“definite tsunami”. About 900 documentary sources, including historical documents, books, scientific reports, newspapers and previous catalogues, support the tsunami data and descriptions gathered in this catalogue. As a result, in the present paper a list of the 290 tsunamis with their main parameters is reported. The online version of the catalogue, available at http://roma2.rm.ingv.it/en/facilities/data_bases/52/catalogue_of_the_euro-mediterranean_tsunamis, provides additional information such as detailed descriptions, pictures, etc. and the complete list of bibliographical sources. Most of the included events have a high reliability value (3= “probable” and 4= “definite” which makes the Euro-Mediterranean Tsunami Catalogue an essential tool for the implementation of tsunami hazard and risk assessment.

  5. Development of ITER PRM and standard parts catalogues in CATIA V5 for tritium-containing systems and components

    International Nuclear Information System (INIS)

    Lazar, Alin; Brad, Sebastian; Zamfirache, Marius; Soare, Sorin; Sofalca, Nicolae; Vijulie, Mihai

    2006-01-01

    CATIA V5 is a software chosen to perform the design and integration within ITER of both systems: fluid and mechanical systems. The broad range of applications provides the ability to develop the design process from the functional 2D design (P and ID) to the 3D plant layouts and detailed design. The 2D symbols for the equipment and piping components were developed in accordance with EN ISO 10628 standard, ISO 3511 (part I, II and IV) standard and DIN 28401 standard and considering the classification and settings done by FZK-team in the PRM-FZK-TLK in order to ensure compliance with the established design standards. These symbols were inserted in 2D catalogues linked with the main catalogue installed in PRM-FZK-TLK on CATIA SERVER at TLK-FZK and intensively tested in Piping and Instrumentation Diagrams. The 3D part for the equipment, instruments and piping components was developed according to the specifications, industrial conventions, terminology and literature available in library, internet and practice. All parts were parametrically built in order to allow the designer to modify the part geometry according to the design. The 3D parts were typed in accordance with IKARUS-Project Instructions by FZK team, inserted in 3D test catalogues and tested in Equipment and Systems workbench. The elements of P and ID, pipes, piping parts, instruments and equipment carry attribute information such as fluid type, pressure rating, insulation or material. The list of attributes is defined in the PRM and consists of the standard industrial attributes implemented by default in CATIA V5 and project specific attributes defined by Project Administration. 2D logical design and 3D Piping models are logically related through the common functions definition in the PRM. The task, TW5-TTFD-TPI-51, was performed collaboratively by MEdC/ICIT and FZK/TLK. (authors)

  6. Overview on the standardization in the field of electromagnetic compatibility

    Science.gov (United States)

    Goldberg, Georges

    1989-04-01

    Standardization in the domain of electromagnetic compatibility (EMC) is discussed, with specific reference to the standards of the International Electrotechnical Commission, the Comite International Special des Perturbations Radioelectriques, and the Comite Europeen de Normalisation Electrotechnique. EMC fields considered include radiocommunications, telecommunications, biological effects, and data transmission. Standards are presented for such electromagnetic disturbances as low-frequency, high-frequency, conduction, and radiation phenomena.

  7. Geospatial semantic web

    CERN Document Server

    Zhang, Chuanrong; Li, Weidong

    2015-01-01

    This book covers key issues related to Geospatial Semantic Web, including geospatial web services for spatial data interoperability; geospatial ontology for semantic interoperability; ontology creation, sharing, and integration; querying knowledge and information from heterogeneous data source; interfaces for Geospatial Semantic Web, VGI (Volunteered Geographic Information) and Geospatial Semantic Web; challenges of Geospatial Semantic Web; and development of Geospatial Semantic Web applications. This book also describes state-of-the-art technologies that attempt to solve these problems such as WFS, WMS, RDF, OWL, and GeoSPARQL, and demonstrates how to use the Geospatial Semantic Web technologies to solve practical real-world problems such as spatial data interoperability.

  8. Catalogue 2.0 the future of the library catalogue

    CERN Document Server

    Chambers, Sally

    2014-01-01

    Brings together some of the foremost international cataloguing practitioners and thought leaders, including Lorcan Dempsey, Emmanuelle Bermès, Marshall Breeding and Karen Calhoun, to provide an overview of the current state of the art of the library catalogue and look ahead to see what the library catalogue might become.

  9. Geospatial Information Service System Based on GeoSOT Grid & Encoding

    Directory of Open Access Journals (Sweden)

    LI Shizhong

    2016-12-01

    Full Text Available With the rapid development of the space and earth observation technology, it is important to establish a multi-source, multi-scale and unified cross-platform reference for global data. In practice, the production and maintenance of geospatial data are scattered in different units, and the standard of the data grid varies between departments and systems. All these bring out the disunity of standards among different historical periods or orgnizations. Aiming at geospatial information security library for the national high resolution earth observation, there are some demands for global display, associated retrieval and template applications and other integrated services for geospatial data. Based on GeoSOT grid and encoding theory system, "geospatial information security library information of globally unified grid encoding management" data subdivision organization solutions have been proposed; system-level analyses, researches and designs have been carried out. The experimental results show that the data organization and management method based on GeoSOT can significantly improve the overall efficiency of the geospatial information security service system.

  10. Catalogue of knowledge and skills for sleep medicine.

    Science.gov (United States)

    Penzel, Thomas; Pevernagie, Dirk; Dogas, Zoran; Grote, Ludger; de Lacy, Simone; Rodenbeck, Andrea; Bassetti, Claudio; Berg, Søren; Cirignotta, Fabio; d'Ortho, Marie-Pia; Garcia-Borreguero, Diego; Levy, Patrick; Nobili, Lino; Paiva, Teresa; Peigneux, Philippe; Pollmächer, Thomas; Riemann, Dieter; Skene, Debra J; Zucconi, Marco; Espie, Colin

    2014-04-01

    Sleep medicine is evolving globally into a medical subspeciality in its own right, and in parallel, behavioural sleep medicine and sleep technology are expanding rapidly. Educational programmes are being implemented at different levels in many European countries. However, these programmes would benefit from a common, interdisciplinary curriculum. This 'catalogue of knowledge and skills' for sleep medicine is proposed, therefore, as a template for developing more standardized curricula across Europe. The Board and The Sleep Medicine Committee of the European Sleep Research Society (ESRS) have compiled the catalogue based on textbooks, standard of practice publications, systematic reviews and professional experience, validated subsequently by an online survey completed by 110 delegates specialized in sleep medicine from different European countries. The catalogue comprises 10 chapters covering physiology, pathology, diagnostic and treatment procedures to societal and organizational aspects of sleep medicine. Required levels of knowledge and skills are defined, as is a proposed workload of 60 points according to the European Credit Transfer System (ECTS). The catalogue is intended to be a basis for sleep medicine education, for sleep medicine courses and for sleep medicine examinations, serving not only physicians with a medical speciality degree, but also PhD and MSc health professionals such as clinical psychologists and scientists, technologists and nurses, all of whom may be involved professionally in sleep medicine. In the future, the catalogue will be revised in accordance with advances in the field of sleep medicine. © 2013 European Sleep Research Society.

  11. BPELPower—A BPEL execution engine for geospatial web services

    Science.gov (United States)

    Yu, Genong (Eugene); Zhao, Peisheng; Di, Liping; Chen, Aijun; Deng, Meixia; Bai, Yuqi

    2012-10-01

    The Business Process Execution Language (BPEL) has become a popular choice for orchestrating and executing workflows in the Web environment. As one special kind of scientific workflow, geospatial Web processing workflows are data-intensive, deal with complex structures in data and geographic features, and execute automatically with limited human intervention. To enable the proper execution and coordination of geospatial workflows, a specially enhanced BPEL execution engine is required. BPELPower was designed, developed, and implemented as a generic BPEL execution engine with enhancements for executing geospatial workflows. The enhancements are especially in its capabilities in handling Geography Markup Language (GML) and standard geospatial Web services, such as the Web Processing Service (WPS) and the Web Feature Service (WFS). BPELPower has been used in several demonstrations over the decade. Two scenarios were discussed in detail to demonstrate the capabilities of BPELPower. That study showed a standard-compliant, Web-based approach for properly supporting geospatial processing, with the only enhancement at the implementation level. Pattern-based evaluation and performance improvement of the engine are discussed: BPELPower directly supports 22 workflow control patterns and 17 workflow data patterns. In the future, the engine will be enhanced with high performance parallel processing and broad Web paradigms.

  12. Arc4nix: A cross-platform geospatial analytical library for cluster and cloud computing

    Science.gov (United States)

    Tang, Jingyin; Matyas, Corene J.

    2018-02-01

    Big Data in geospatial technology is a grand challenge for processing capacity. The ability to use a GIS for geospatial analysis on Cloud Computing and High Performance Computing (HPC) clusters has emerged as a new approach to provide feasible solutions. However, users lack the ability to migrate existing research tools to a Cloud Computing or HPC-based environment because of the incompatibility of the market-dominating ArcGIS software stack and Linux operating system. This manuscript details a cross-platform geospatial library "arc4nix" to bridge this gap. Arc4nix provides an application programming interface compatible with ArcGIS and its Python library "arcpy". Arc4nix uses a decoupled client-server architecture that permits geospatial analytical functions to run on the remote server and other functions to run on the native Python environment. It uses functional programming and meta-programming language to dynamically construct Python codes containing actual geospatial calculations, send them to a server and retrieve results. Arc4nix allows users to employ their arcpy-based script in a Cloud Computing and HPC environment with minimal or no modification. It also supports parallelizing tasks using multiple CPU cores and nodes for large-scale analyses. A case study of geospatial processing of a numerical weather model's output shows that arcpy scales linearly in a distributed environment. Arc4nix is open-source software.

  13. Global polar geospatial information service retrieval based on search engine and ontology reasoning

    Science.gov (United States)

    Chen, Nengcheng; E, Dongcheng; Di, Liping; Gong, Jianya; Chen, Zeqiang

    2007-01-01

    In order to improve the access precision of polar geospatial information service on web, a new methodology for retrieving global spatial information services based on geospatial service search and ontology reasoning is proposed, the geospatial service search is implemented to find the coarse service from web, the ontology reasoning is designed to find the refined service from the coarse service. The proposed framework includes standardized distributed geospatial web services, a geospatial service search engine, an extended UDDI registry, and a multi-protocol geospatial information service client. Some key technologies addressed include service discovery based on search engine and service ontology modeling and reasoning in the Antarctic geospatial context. Finally, an Antarctica multi protocol OWS portal prototype based on the proposed methodology is introduced.

  14. NCI's Distributed Geospatial Data Server

    Science.gov (United States)

    Larraondo, P. R.; Evans, B. J. K.; Antony, J.

    2016-12-01

    Earth systems, environmental and geophysics datasets are an extremely valuable source of information about the state and evolution of the Earth. However, different disciplines and applications require this data to be post-processed in different ways before it can be used. For researchers experimenting with algorithms across large datasets or combining multiple data sets, the traditional approach to batch data processing and storing all the output for later analysis rapidly becomes unfeasible, and often requires additional work to publish for others to use. Recent developments on distributed computing using interactive access to significant cloud infrastructure opens the door for new ways of processing data on demand, hence alleviating the need for storage space for each individual copy of each product. The Australian National Computational Infrastructure (NCI) has developed a highly distributed geospatial data server which supports interactive processing of large geospatial data products, including satellite Earth Observation data and global model data, using flexible user-defined functions. This system dynamically and efficiently distributes the required computations among cloud nodes and thus provides a scalable analysis capability. In many cases this completely alleviates the need to preprocess and store the data as products. This system presents a standards-compliant interface, allowing ready accessibility for users of the data. Typical data wrangling problems such as handling different file formats and data types, or harmonising the coordinate projections or temporal and spatial resolutions, can now be handled automatically by this service. The geospatial data server exposes functionality for specifying how the data should be aggregated and transformed. The resulting products can be served using several standards such as the Open Geospatial Consortium's (OGC) Web Map Service (WMS) or Web Feature Service (WFS), Open Street Map tiles, or raw binary arrays under

  15. Developing a prenatal nursing care International Classification for Nursing Practice catalogue.

    Science.gov (United States)

    Liu, L; Coenen, A; Tao, H; Jansen, K R; Jiang, A L

    2017-09-01

    This study aimed to develop a prenatal nursing care catalogue of International Classification for Nursing Practice. As a programme of the International Council of Nurses, International Classification for Nursing Practice aims to support standardized electronic nursing documentation and facilitate collection of comparable nursing data across settings. This initiative enables the study of relationships among nursing diagnoses, nursing interventions and nursing outcomes for best practice, healthcare management decisions, and policy development. The catalogues are usually focused on target populations. Pregnant women are the nursing population addressed in this project. According to the guidelines for catalogue development, three research steps have been adopted: (a) identifying relevant nursing diagnoses, interventions and outcomes; (b) developing a conceptual framework for the catalogue; (c) expert's validation. This project established a prenatal nursing care catalogue with 228 terms in total, including 69 nursing diagnosis, 92 nursing interventions and 67 nursing outcomes, among them, 57 nursing terms were newly developed. All terms in the catalogue were organized by a framework with two main categories, i.e. Expected Changes of Pregnancy and Pregnancy at Risk. Each category had four domains, representing the physical, psychological, behavioral and environmental perspectives of nursing practice. This catalogue can ease the documentation workload among prenatal care nurses, and facilitate storage and retrieval of standardized data for many purposes, such as quality improvement, administration decision-support and researches. The documentations of prenatal care provided data that can be more fluently communicated, compared and evaluated across various healthcare providers and clinic settings. © 2016 International Council of Nurses.

  16. Towards Precise Metadata-set for Discovering 3D Geospatial Models in Geo-portals

    Science.gov (United States)

    Zamyadi, A.; Pouliot, J.; Bédard, Y.

    2013-09-01

    Accessing 3D geospatial models, eventually at no cost and for unrestricted use, is certainly an important issue as they become popular among participatory communities, consultants, and officials. Various geo-portals, mainly established for 2D resources, have tried to provide access to existing 3D resources such as digital elevation model, LIDAR or classic topographic data. Describing the content of data, metadata is a key component of data discovery in geo-portals. An inventory of seven online geo-portals and commercial catalogues shows that the metadata referring to 3D information is very different from one geo-portal to another as well as for similar 3D resources in the same geo-portal. The inventory considered 971 data resources affiliated with elevation. 51% of them were from three geo-portals running at Canadian federal and municipal levels whose metadata resources did not consider 3D model by any definition. Regarding the remaining 49% which refer to 3D models, different definition of terms and metadata were found, resulting in confusion and misinterpretation. The overall assessment of these geo-portals clearly shows that the provided metadata do not integrate specific and common information about 3D geospatial models. Accordingly, the main objective of this research is to improve 3D geospatial model discovery in geo-portals by adding a specific metadata-set. Based on the knowledge and current practices on 3D modeling, and 3D data acquisition and management, a set of metadata is proposed to increase its suitability for 3D geospatial models. This metadata-set enables the definition of genuine classes, fields, and code-lists for a 3D metadata profile. The main structure of the proposal contains 21 metadata classes. These classes are classified in three packages as General and Complementary on contextual and structural information, and Availability on the transition from storage to delivery format. The proposed metadata set is compared with Canadian Geospatial

  17. National Geospatial Program

    Science.gov (United States)

    Carswell, William J.

    2011-01-01

    The National Geospatial Program (NGP; http://www.usgs.gov/ngpo/) satisfies the needs of customers by providing geospatial products and services that customers incorporate into their decisionmaking and operational activities. These products and services provide geospatial data that are organized and maintained in cost-effective ways and developed by working with partners and organizations whose activities align with those of the program. To accomplish its mission, the NGP— organizes, maintains, publishes, and disseminates the geospatial baseline of the Nation's topography, natural landscape, and manmade environment through The National Map

  18. User-driven generation of standard data services

    Science.gov (United States)

    Díaz, Laura; Granell, Carlos; Gould, Michael; Huerta, Joaquín.

    2010-05-01

    generation and publication of standard geospatial are proposed. These mechanisms deal with the fact of hiding the underlying technology and let stakeholders wrap resources as standard services to share these resources in a transparent manner. These components are integrated in our architecture within the Service Framework node (module). PIC Figure 1: Figure 1. Architecture components diagram Figure 1 shows the components of the architecture: The Application Node provides the entry point for users to run distributed applications. This software component has the user interface and the application logic. The Service Connector component provides the ability to connect to the services available in the middleware layer of SDI. This node acts as a socket to OGC Web Services. For instance we appreciate the WMS component implementing the OGC WMS specification as it is the standard recommended by the INSPIRE implementation rules as View Service Type.The Service Framework node contains several components. The Service Framework main functionality is to assist users in wrapping and sharing geospatial resources. It implements the proposed mechanisms to improve the availability and visibility of geospatial resources. The main components of this framework are the Data wrapper, the Process Wrapper and the Service Publisher. The Data Wrapper and Process Wrapper components guide users to wrap data and tools as standard services according with INSPIRE implementing rules (availability). The Service Publisher component aims at creating service metadata and publishing them in catalogues (visibility). Roughly speaking, all of these components are concerned with the idea of acting as a service generator and publisher, i.e., they get a resource (data or process) and return an INSPIRE service that will be published in catalogue services. References Béjar, R., Latre, M. Á., Nogueras-Iso, J., Muro-Medrano, P. R., Zarazaga-Soria, F. J. 2009. International Journal of Geographical Information Science, 23

  19. A Javascript GIS Platform Based on Invocable Geospatial Web Services

    Directory of Open Access Journals (Sweden)

    Konstantinos Evangelidis

    2018-04-01

    Full Text Available Semantic Web technologies are being increasingly adopted by the geospatial community during last decade through the utilization of open standards for expressing and serving geospatial data. This was also dramatically assisted by the ever-increasing access and usage of geographic mapping and location-based services via smart devices in people’s daily activities. In this paper, we explore the developmental framework of a pure JavaScript client-side GIS platform exclusively based on invocable geospatial Web services. We also extend JavaScript utilization on the server side by deploying a node server acting as a bridge between open source WPS libraries and popular geoprocessing engines. The vehicle for such an exploration is a cross platform Web browser capable of interpreting JavaScript commands to achieve interaction with geospatial providers. The tool is a generic Web interface providing capabilities of acquiring spatial datasets, composing layouts and applying geospatial processes. In an ideal form the end-user will have to identify those services, which satisfy a geo-related need and put them in the appropriate row. The final output may act as a potential collector of freely available geospatial web services. Its server-side components may exploit geospatial processing suppliers composing that way a light-weight fully transparent open Web GIS platform.

  20. Enhancing discovery in spatial data infrastructures using a search engine

    Directory of Open Access Journals (Sweden)

    Paolo Corti

    2018-05-01

    Full Text Available A spatial data infrastructure (SDI is a framework of geospatial data, metadata, users and tools intended to provide an efficient and flexible way to use spatial information. One of the key software components of an SDI is the catalogue service which is needed to discover, query and manage the metadata. Catalogue services in an SDI are typically based on the Open Geospatial Consortium (OGC Catalogue Service for the Web (CSW standard which defines common interfaces for accessing the metadata information. A search engine is a software system capable of supporting fast and reliable search, which may use ‘any means necessary’ to get users to the resources they need quickly and efficiently. These techniques may include full text search, natural language processing, weighted results, fuzzy tolerance results, faceting, hit highlighting, recommendations and many others. In this paper we present an example of a search engine being added to an SDI to improve search against large collections of geospatial datasets. The Centre for Geographic Analysis (CGA at Harvard University re-engineered the search component of its public domain SDI (Harvard WorldMap which is based on the GeoNode platform. A search engine was added to the SDI stack to enhance the CSW catalogue discovery abilities. It is now possible to discover spatial datasets from metadata by using the standard search operations of the catalogue and to take advantage of the new abilities of the search engine, to return relevant and reliable content to SDI users.

  1. Users and Union Catalogues

    Science.gov (United States)

    Hartley, R. J.; Booth, Helen

    2006-01-01

    Union catalogues have had an important place in libraries for many years. Their use has been little investigated. Recent interest in the relative merits of physical and virtual union catalogues and a recent collaborative project between a physical and several virtual union catalogues in the United Kingdom led to the opportunity to study how users…

  2. Geospatial Brokering - Challenges and Future Directions

    Science.gov (United States)

    White, C. E.

    2012-12-01

    An important feature of many brokers is to facilitate straightforward human access to scientific data while maintaining programmatic access to it for system solutions. Standards-based protocols are critical for this, and there are a number of protocols to choose from. In this discussion, we will present a web application solution that leverages certain protocols - e.g., OGC CSW, REST, and OpenSearch - to provide programmatic as well as human access to geospatial resources. We will also discuss managing resources to reduce duplication yet increase discoverability, federated search solutions, and architectures that combine human-friendly interfaces with powerful underlying data management. The changing requirements witnessed in brokering solutions over time, our recent experience participating in the EarthCube brokering hack-a-thon, and evolving interoperability standards provide insight to future technological and philosophical directions planned for geospatial broker solutions. There has been much change over the past decade, but with the unprecedented data collaboration of recent years, in many ways the challenges and opportunities are just beginning.

  3. Geospatial Services Laboratory

    Data.gov (United States)

    Federal Laboratory Consortium — FUNCTION: To process, store, and disseminate geospatial data to the Department of Defense and other Federal agencies.DESCRIPTION: The Geospatial Services Laboratory...

  4. Multifractal Omori law for earthquake triggering: new tests on the California, Japan and worldwide catalogues

    Science.gov (United States)

    Ouillon, G.; Sornette, D.; Ribeiro, E.

    2009-07-01

    The Multifractal Stress-Activated model is a statistical model of triggered seismicity based on mechanical and thermodynamic principles. It predicts that, above a triggering magnitude cut-off M0, the exponent p of the Omori law for the time decay of the rate of aftershocks is a linear increasing function p(M) = a0M + b0 of the main shock magnitude M. We previously reported empirical support for this prediction, using the Southern California Earthquake Center (SCEC) catalogue. Here, we confirm this observation using an updated, longer version of the same catalogue, as well as new methods to estimate p. One of this methods is the newly defined Scaling Function Analysis (SFA), adapted from the wavelet transform. This method is able to measure a mathematical singularity (hence a p-value), erasing the possible regular part of a time-series. The SFA also proves particularly efficient to reveal the coexistence and superposition of several types of relaxation laws (typical Omori sequences and short-lived swarms sequences) which can be mixed within the same catalogue. Another new method consists in monitoring the largest aftershock magnitude observed in successive time intervals, and thus shortcuts the problem of missing events with small magnitudes in aftershock catalogues. The same methods are used on data from the worldwide Harvard Centroid Moment Tensor (CMT) catalogue and show results compatible with those of Southern California. For the Japan Meteorological Agency (JMA) catalogue, we still observe a linear dependence of p on M, but with a smaller slope. The SFA shows however that results for this catalogue may be biased by numerous swarm sequences, despite our efforts to remove them before the analysis.

  5. Remapping dark matter halo catalogues between cosmological simulations

    Science.gov (United States)

    Mead, A. J.; Peacock, J. A.

    2014-05-01

    We present and test a method for modifying the catalogue of dark matter haloes produced from a given cosmological simulation, so that it resembles the result of a simulation with an entirely different set of parameters. This extends the method of Angulo & White, which rescales the full particle distribution from a simulation. Working directly with the halo catalogue offers an advantage in speed, and also allows modifications of the internal structure of the haloes to account for non-linear differences between cosmologies. Our method can be used directly on a halo catalogue in a self-contained manner without any additional information about the overall density field; although the large-scale displacement field is required by the method, this can be inferred from the halo catalogue alone. We show proof of concept of our method by rescaling a matter-only simulation with no baryon acoustic oscillation (BAO) features to a more standard Λ cold dark matter model containing a cosmological constant and a BAO signal. In conjunction with the halo occupation approach, this method provides a basis for the rapid generation of mock galaxy samples spanning a wide range of cosmological parameters.

  6. The IKEA Catalogue

    DEFF Research Database (Denmark)

    Brown, Barry; Bleecker, Julian; D'Adamo, Marco

    2016-01-01

    This paper is an introduction to the "Future IKEA Catalogue", enclosed here as an example of a design fiction produced from a long standing industrial-academic collaboration. We introduce the catalogue here by discussing some of our experiences using design fiction` with companies and public sector...

  7. Addressing the Challenge: Cataloguing Electronic Books in Academic Libraries

    Directory of Open Access Journals (Sweden)

    Shuzhen Zhao

    2010-03-01

    Full Text Available Objective ‐ This paper explores the various issues and challenges arising from e‐book cataloguing experienced at the University of Windsor’s Leddy Library and the Ontario Council of University Libraries (OCUL. This discussion uses an evidence based approach to identify and resolve issues relevant to academic libraries as well as to consortia. With the ever rising popularity of e‐books within academic libraries, cataloguing librarians are actively seeking more effective methods of managing this new electronic medium, including the development of new cataloguing policies and procedures. This paper will explore the various issues and challenges surrounding e‐book cataloguing and processing within academic libraries, and will identify new policies and procedures that may be used to effectively assist in e‐book management.Methods ‐ This paper presents a case study of e‐book cataloguing practices undertaken by a Canadian academic library and the consortium with which it is affiliated. Towards this end, the University of Windsor’s Leddy Library will be the prime focus of this study, with its establishment of a new e‐book MARC records database. The research is based on the results of the e‐book MARC project undertaken by the Leddy Library and the Ontario Council of University Libraries (OCUL.Through analysis of various suppliers’ MARC records and the actual implementation of the e‐book MARC project, the authors developed and evaluated a new approach to e‐book cataloguing for use in academic libraries.Results ‐ This practice‐based approach towards the development of a new method of e‐book cataloguing required continual modification and examination of e‐book MARC records within the target library. The Leddy Library’s e‐book MARC project provided an excellent opportunity to test the library’s existing cataloguing standards and procedures for print format, while at the same time, identifying related e‐books issues

  8. Cataloguing outside the box a practical guide to cataloguing special collections materials

    CERN Document Server

    Falk, Patricia

    2010-01-01

    A practical guide to cataloguing and processing the unique special collections formats in the Browne Popular Culture Library (BPCL) and the Music Library and Sound Recordings Archives (MLSRA) at Bowling Green State University (BGSU) (e.g. fanzines, popular sound recordings, comic books, motion picture scripts and press kits, popular fiction). Cataloguing Outside the Box provides guidance to professionals in library and information science facing the same cataloguing challenges. Additionally, name authority work for these collections is addressed.provides practical guidelines and solutions for

  9. INDXENDF, Preparation of Visual Catalogue of ENDF Format Data

    International Nuclear Information System (INIS)

    Silva, Orion de O.; Paviotti Corcuera, R.; De Moraes Cunha, M.; Ferreira, P.A.

    1996-01-01

    1 - Description of program or function: This program is a video catalogue for libraries in the ENDF-4, ENDF-5 or ENDF-6 format (Evaluated Nuclear Data File) which can be run on an IBM-PC or compatible computer. This user friendly catalogue is of interest to nuclear and reactor physics researchers. The input is the filename of ENDF data and the two output files contain: i. the list of materials with corresponding laboratory, author and date of evaluation; ii. information about the MF and MT numbers for each material. The program is written in the C language whose capability of providing windows and 'interrupts' along with speed and portability, has been greatly exploited. The system allows output of options (i) and (ii) either on screen, printer or hard disk. 2 - Method of solution: The source code of about 3000 lines was written in C. The routines for windowing were based on the following works: Hummel (1988), Stevens (1989), Lafore (1987), Borland International (1988a, 1988b) and Schildt (1988, 1989). 3 - Restrictions on the complexity of the problem: The executable program occupies about 52 Kb of memory. The extra hard disk space needed depends upon the size of the ENDF/B data file to be processed (e.g. the Activation file contains about 1.3 M-bytes, the General Purpose ENDF/B-VI has four parts, each containing about 12 M-bytes). To run the program the 'datafile' and the executable code '.EXE' file should be on the hard-drive. The program may be run on any IBM/PC or compatible with at least 640 Kb RAM

  10. On Hydronymic Catalogues Composition Principles: Cataloguing of Hydronyms of the Msta River Basin

    OpenAIRE

    Valery L. Vasilyev; Nina N. Vikhrova

    2015-01-01

    The article presents a brief review of the few Russian hydronymic catalogues (relating to the basins of the Don, Oka, Svir and other rivers) based on the hydrographic principle. The authors argue that, in comparison with alphabetized hydronymic dictionaries, hydronymic catalogues have some obvious advantages for onomastic lexicography. This kind of catalogues should include, firstly, all historically attested forms of a hydronym (including those considered to be occasional miswritings) and, s...

  11. A FRAMEWORK FOR AN OPEN SOURCE GEOSPATIAL CERTIFICATION MODEL

    Directory of Open Access Journals (Sweden)

    T. U. R. Khan

    2016-06-01

    Full Text Available The geospatial industry is forecasted to have an enormous growth in the forthcoming years and an extended need for well-educated workforce. Hence ongoing education and training play an important role in the professional life. Parallel, in the geospatial and IT arena as well in the political discussion and legislation Open Source solutions, open data proliferation, and the use of open standards have an increasing significance. Based on the Memorandum of Understanding between International Cartographic Association, OSGeo Foundation, and ISPRS this development led to the implementation of the ICA-OSGeo-Lab imitative with its mission “Making geospatial education and opportunities accessible to all”. Discussions in this initiative and the growth and maturity of geospatial Open Source software initiated the idea to develop a framework for a worldwide applicable Open Source certification approach. Generic and geospatial certification approaches are already offered by numerous organisations, i.e., GIS Certification Institute, GeoAcademy, ASPRS, and software vendors, i. e., Esri, Oracle, and RedHat. They focus different fields of expertise and have different levels and ways of examination which are offered for a wide range of fees. The development of the certification framework presented here is based on the analysis of diverse bodies of knowledge concepts, i.e., NCGIA Core Curriculum, URISA Body Of Knowledge, USGIF Essential Body Of Knowledge, the “Geographic Information: Need to Know", currently under development, and the Geospatial Technology Competency Model (GTCM. The latter provides a US American oriented list of the knowledge, skills, and abilities required of workers in the geospatial technology industry and influenced essentially the framework of certification. In addition to the theoretical analysis of existing resources the geospatial community was integrated twofold. An online survey about the relevance of Open Source was performed and

  12. a Framework for AN Open Source Geospatial Certification Model

    Science.gov (United States)

    Khan, T. U. R.; Davis, P.; Behr, F.-J.

    2016-06-01

    The geospatial industry is forecasted to have an enormous growth in the forthcoming years and an extended need for well-educated workforce. Hence ongoing education and training play an important role in the professional life. Parallel, in the geospatial and IT arena as well in the political discussion and legislation Open Source solutions, open data proliferation, and the use of open standards have an increasing significance. Based on the Memorandum of Understanding between International Cartographic Association, OSGeo Foundation, and ISPRS this development led to the implementation of the ICA-OSGeo-Lab imitative with its mission "Making geospatial education and opportunities accessible to all". Discussions in this initiative and the growth and maturity of geospatial Open Source software initiated the idea to develop a framework for a worldwide applicable Open Source certification approach. Generic and geospatial certification approaches are already offered by numerous organisations, i.e., GIS Certification Institute, GeoAcademy, ASPRS, and software vendors, i. e., Esri, Oracle, and RedHat. They focus different fields of expertise and have different levels and ways of examination which are offered for a wide range of fees. The development of the certification framework presented here is based on the analysis of diverse bodies of knowledge concepts, i.e., NCGIA Core Curriculum, URISA Body Of Knowledge, USGIF Essential Body Of Knowledge, the "Geographic Information: Need to Know", currently under development, and the Geospatial Technology Competency Model (GTCM). The latter provides a US American oriented list of the knowledge, skills, and abilities required of workers in the geospatial technology industry and influenced essentially the framework of certification. In addition to the theoretical analysis of existing resources the geospatial community was integrated twofold. An online survey about the relevance of Open Source was performed and evaluated with 105

  13. On Hydronymic Catalogues Composition Principles: Cataloguing of Hydronyms of the Msta River Basin

    Directory of Open Access Journals (Sweden)

    Valery L. Vasilyev

    2015-06-01

    Full Text Available The article presents a brief review of the few Russian hydronymic catalogues (relating to the basins of the Don, Oka, Svir and other rivers based on the hydrographic principle. The authors argue that, in comparison with alphabetized hydronymic dictionaries, hydronymic catalogues have some obvious advantages for onomastic lexicography. This kind of catalogues should include, firstly, all historically attested forms of a hydronym (including those considered to be occasional miswritings and, secondly, all non-hydronymic names making part of the respective hydronymic microsystem and providing “external” (i. e., chronological, derivational, etymological, ethno-historical information about the hydronym. The authors point out that the cataloguing of hydronyms based on the hydrographic principle entails some difficulties: impossibility to localize some bodies of water mentioned in ancient and medieval documents; differences in the indication of the same bodies of water on old and contemporary maps; historical differences in establishing hydrographic hierarchies; historical changes of lake-river systems, etc. The authors also share their experience in creating a hydronymic catalogue of the Msta River basin in Novgorod and Tver Regions of Russia. They describe the principles of the composition of the catalogue and present a short excerpt of it that orders names in the system of the Volma River, one of the Msta’s left tributaries.

  14. Python geospatial development

    CERN Document Server

    Westra, Erik

    2013-01-01

    This is a tutorial style book that will teach usage of Python tools for GIS using simple practical examples and then show you how to build a complete mapping application from scratch. The book assumes basic knowledge of Python. No knowledge of Open Source GIS is required.Experienced Python developers who want to learn about geospatial concepts, work with geospatial data, solve spatial problems, and build mapbased applications.This book will be useful those who want to get up to speed with Open Source GIS in order to build GIS applications or integrate GeoSpatial features into their existing ap

  15. Geospatial Data as a Service: Towards planetary scale real-time analytics

    Science.gov (United States)

    Evans, B. J. K.; Larraondo, P. R.; Antony, J.; Richards, C. J.

    2017-12-01

    The rapid growth of earth systems, environmental and geophysical datasets poses a challenge to both end-users and infrastructure providers. For infrastructure and data providers, tasks like managing, indexing and storing large collections of geospatial data needs to take into consideration the various use cases by which consumers will want to access and use the data. Considerable investment has been made by the Earth Science community to produce suitable real-time analytics platforms for geospatial data. There are currently different interfaces that have been defined to provide data services. Unfortunately, there is considerable difference on the standards, protocols or data models which have been designed to target specific communities or working groups. The Australian National University's National Computational Infrastructure (NCI) is used for a wide range of activities in the geospatial community. Earth observations, climate and weather forecasting are examples of these communities which generate large amounts of geospatial data. The NCI has been carrying out significant effort to develop a data and services model that enables the cross-disciplinary use of data. Recent developments in cloud and distributed computing provide a publicly accessible platform where new infrastructures can be built. One of the key components these technologies offer is the possibility of having "limitless" compute power next to where the data is stored. This model is rapidly transforming data delivery from centralised monolithic services towards ubiquitous distributed services that scale up and down adapting to fluctuations in the demand. NCI has developed GSKY, a scalable, distributed server which presents a new approach for geospatial data discovery and delivery based on OGC standards. We will present the architecture and motivating use-cases that drove GSKY's collaborative design, development and production deployment. We show our approach offers the community valuable exploratory

  16. Geospatial Authentication

    Science.gov (United States)

    Lyle, Stacey D.

    2009-01-01

    A software package that has been designed to allow authentication for determining if the rover(s) is/are within a set of boundaries or a specific area to access critical geospatial information by using GPS signal structures as a means to authenticate mobile devices into a network wirelessly and in real-time. The advantage lies in that the system only allows those with designated geospatial boundaries or areas into the server.

  17. Infrastructure for the Geospatial Web

    Science.gov (United States)

    Lake, Ron; Farley, Jim

    Geospatial data and geoprocessing techniques are now directly linked to business processes in many areas. Commerce, transportation and logistics, planning, defense, emergency response, health care, asset management and many other domains leverage geospatial information and the ability to model these data to achieve increased efficiencies and to develop better, more comprehensive decisions. However, the ability to deliver geospatial data and the capacity to process geospatial information effectively in these domains are dependent on infrastructure technology that facilitates basic operations such as locating data, publishing data, keeping data current and notifying subscribers and others whose applications and decisions are dependent on this information when changes are made. This chapter introduces the notion of infrastructure technology for the Geospatial Web. Specifically, the Geography Markup Language (GML) and registry technology developed using the ebRIM specification delivered from the OASIS consortium are presented as atomic infrastructure components in a working Geospatial Web.

  18. Geospatial Semantics and the Semantic Web

    CERN Document Server

    Ashish, Naveen

    2011-01-01

    The availability of geographic and geospatial information and services, especially on the open Web has become abundant in the last several years with the proliferation of online maps, geo-coding services, geospatial Web services and geospatially enabled applications. The need for geospatial reasoning has significantly increased in many everyday applications including personal digital assistants, Web search applications, local aware mobile services, specialized systems for emergency response, medical triaging, intelligence analysis and more. Geospatial Semantics and the Semantic Web: Foundation

  19. SMI Compatible Simulation Scheduler Design for Reuse of Model Complying with SMP Standard

    Directory of Open Access Journals (Sweden)

    Cheol-Hea Koo

    2010-12-01

    Full Text Available Software reusability is one of key factors which impacts cost and schedule on a software development project. It is very crucial also in satellite simulator development since there are many commercial simulator models related to satellite and dynamics. If these models can be used in another simulator platform, great deal of confidence and cost/schedule reduction would be achieved. Simulation model portability (SMP is maintained by European Space Agency and many models compatible with SMP/simulation model interface (SMI are available. Korea Aerospace Research Institute (KARI is developing hardware abstraction layer (HAL supported satellite simulator to verify on-board software of satellite. From above reasons, KARI wants to port these SMI compatible models to the HAL supported satellite simulator. To port these SMI compatible models to the HAL supported satellite simulator, simulation scheduler is preliminary designed according to the SMI standard.

  20. The two union catalogues of Myanmar

    Energy Technology Data Exchange (ETDEWEB)

    Hla, Win [Myanmar Scientific and Technological Research Dept., Yangon (Myanmar)

    1995-04-01

    The article mentions about the two union catalogues of Myanmar. The first one is the ``Consolidated Catalogue of journals and the periodicals contained in the libraries of Kasuali, Calcutta, Bombay, Madras, Coonoor, Rangoon and Shillong``. This was published by Indian Research Fund Association of Calcutta in 1933. This is the first union catalogue of medical periodicals for both Myanmar and India as well. The second one is ``the Regional Union Catalogue of Scientific Serials: Yangon``. This was published in 1977, its second printing in 1989. This union catalogue excludes medical serials. Twenty libraries took part in the compilation and publishing of the union catalogue with Technical Information Centre of Myanmar Scientific and Technological Research Department, (formerly Central Research Organization), No. 6, Kaba Aye Pagoda Road, Yankin P.O. Yangon, Myanmar, taking the leading role.

  1. The two union catalogues of Myanmar

    International Nuclear Information System (INIS)

    Hla, Win

    1995-01-01

    The article mentions about the two union catalogues of Myanmar. The first one is the ''Consolidated Catalogue of journals and the periodicals contained in the libraries of Kasuali, Calcutta, Bombay, Madras, Coonoor, Rangoon and Shillong''. This was published by Indian Research Fund Association of Calcutta in 1933. This is the first union catalogue of medical periodicals for both Myanmar and India as well. The second one is ''the Regional Union Catalogue of Scientific Serials: Yangon''. This was published in 1977, its second printing in 1989. This union catalogue excludes medical serials. Twenty libraries took part in the compilation and publishing of the union catalogue with Technical Information Centre of Myanmar Scientific and Technological Research Department, (formerly Central Research Organization), No. 6, Kaba Aye Pagoda Road, Yankin P.O. Yangon, Myanmar, taking the leading role

  2. The Availability of MeSH in Vendor-Supplied Cataloguing Records, as Seen Through the Catalogue of a Canadian Academic Health Library

    Directory of Open Access Journals (Sweden)

    Pamela S. Morgan

    2007-09-01

    Full Text Available This study examines the prevalence of medical subject headings in vendor-supplied cataloguing records for publications contained within aggregated databases or publisher collections. In the first phase, the catalogue of one Canadian academic medical library was examined to determine the extent to which medical subject headings (MeSH are available in the vendor-supplied records. In the second phase, these results were compared to the catalogues of other Canadian academic medical libraries in order to reach a generalization regarding the availability of MeSH headings for electronic resources. MeSH was more widespread in records for electronic journals but was noticeably lacking in records for electronic monographs, and for Canadian publications. There is no standard for ensuring MeSH are assigned to monograph records for health titles and there is no library in Canada with responsibility for ensuring that Canadian health publications receive Medical Subject Headings. It is incumbent upon libraries using MeSH to ensure that vendors are aware of this need when purchasing record sets.

  3. Geospatial health

    DEFF Research Database (Denmark)

    Utzinger, Jürg; Rinaldi, Laura; Malone, John B.

    2011-01-01

    Geospatial Health is an international, peer-reviewed scientific journal produced by the Global Network for Geospatial Health (GnosisGIS). This network was founded in 2000 and the inaugural issue of its official journal was published in November 2006 with the aim to cover all aspects of geographical...... information system (GIS) applications, remote sensing and other spatial analytic tools focusing on human and veterinary health. The University of Naples Federico II is the publisher, producing two issues per year, both as hard copy and an open-access online version. The journal is referenced in major...... databases, including CABI, ISI Web of Knowledge and PubMed. In 2008, it was assigned its first impact factor (1.47), which has now reached 1.71. Geospatial Health is managed by an editor-in-chief and two associate editors, supported by five regional editors and a 23-member strong editorial board...

  4. Emerging Geospatial Sharing Technologies in Earth and Space Science Informatics

    Science.gov (United States)

    Singh, R.; Bermudez, L. E.

    2013-12-01

    Emerging Geospatial Sharing Technologies in Earth and Space Science Informatics The Open Geospatial Consortium (OGC) mission is to serve as a global forum for the collaboration of developers and users of spatial data products and services, and to advance the development of international standards for geospatial interoperability. The OGC coordinates with over 400 institutions in the development of geospatial standards. In the last years two main trends are making disruptions in geospatial applications: mobile and context sharing. People now have more and more mobile devices to support their work and personal life. Mobile devices are intermittently connected to the internet and have smaller computing capacity than a desktop computer. Based on this trend a new OGC file format standard called GeoPackage will enable greater geospatial data sharing on mobile devices. GeoPackage is perhaps best understood as the natural evolution of Shapefiles, which have been the predominant lightweight geodata sharing format for two decades. However the format is extremely limited. Four major shortcomings are that only vector points, lines, and polygons are supported; property names are constrained by the dBASE format; multiple files are required to encode a single data set; and multiple Shapefiles are required to encode multiple data sets. A more modern lingua franca for geospatial data is long overdue. GeoPackage fills this need with support for vector data, image tile matrices, and raster data. And it builds upon a database container - SQLite - that's self-contained, single-file, cross-platform, serverless, transactional, and open source. A GeoPackage, in essence, is a set of SQLite database tables whose content and layout is described in the candidate GeoPackage Implementation Specification available at https://portal.opengeospatial.org/files/?artifact_id=54838&version=1. The second trend is sharing client 'contexts'. When a user is looking into an article or a product on the web

  5. The African Geospatial Sciences Institute (agsi): a New Approach to Geospatial Training in North Africa

    Science.gov (United States)

    Oeldenberger, S.; Khaled, K. B.

    2012-07-01

    The African Geospatial Sciences Institute (AGSI) is currently being established in Tunisia as a non-profit, non-governmental organization (NGO). Its objective is to accelerate the geospatial capacity development in North-Africa, providing the facilities for geospatial project and management training to regional government employees, university graduates, private individuals and companies. With typical course durations between one and six months, including part-time programs and long-term mentoring, its focus is on practical training, providing actual project execution experience. The AGSI will complement formal university education and will work closely with geospatial certification organizations and the geospatial industry. In the context of closer cooperation between neighboring North Africa and the European Community, the AGSI will be embedded in a network of several participating European and African universities, e. g. the ITC, and international organizations, such as the ISPRS, the ICA and the OGC. Through a close cooperation with African organizations, such as the AARSE, the RCMRD and RECTAS, the network and exchange of ideas, experiences, technology and capabilities will be extended to Saharan and sub-Saharan Africa. A board of trustees will be steering the AGSI operations and will ensure that practical training concepts and contents are certifiable and can be applied within a credit system to graduate and post-graduate education at European and African universities. The geospatial training activities of the AGSI are centered on a facility with approximately 30 part- and full-time general staff and lecturers in Tunis during the first year. The AGSI will operate a small aircraft with a medium-format aerial camera and compact LIDAR instrument for local, community-scale data capture. Surveying training, the photogrammetric processing of aerial images, GIS data capture and remote sensing training will be the main components of the practical training courses

  6. Geospatial Data Management Platform for Urban Groundwater

    Science.gov (United States)

    Gaitanaru, D.; Priceputu, A.; Gogu, C. R.

    2012-04-01

    Due to the large amount of civil work projects and research studies, large quantities of geo-data are produced for the urban environments. These data are usually redundant as well as they are spread in different institutions or private companies. Time consuming operations like data processing and information harmonisation represents the main reason to systematically avoid the re-use of data. The urban groundwater data shows the same complex situation. The underground structures (subway lines, deep foundations, underground parkings, and others), the urban facility networks (sewer systems, water supply networks, heating conduits, etc), the drainage systems, the surface water works and many others modify continuously. As consequence, their influence on groundwater changes systematically. However, these activities provide a large quantity of data, aquifers modelling and then behaviour prediction can be done using monitored quantitative and qualitative parameters. Due to the rapid evolution of technology in the past few years, transferring large amounts of information through internet has now become a feasible solution for sharing geoscience data. Furthermore, standard platform-independent means to do this have been developed (specific mark-up languages like: GML, GeoSciML, WaterML, GWML, CityML). They allow easily large geospatial databases updating and sharing through internet, even between different companies or between research centres that do not necessarily use the same database structures. For Bucharest City (Romania) an integrated platform for groundwater geospatial data management is developed under the framework of a national research project - "Sedimentary media modeling platform for groundwater management in urban areas" (SIMPA) financed by the National Authority for Scientific Research of Romania. The platform architecture is based on three components: a geospatial database, a desktop application (a complex set of hydrogeological and geological analysis

  7. Standardized, utility-DOE compatible, spent fuel storage-transport systems

    International Nuclear Information System (INIS)

    Smith, M.L.

    1991-01-01

    Virginia Power has developed and licensed a facility for dry storage of spent nuclear fuel in metal spent fuel storage casks. The modifications to the design of these casks necessary for licensing for both storage and transport of spent fuel are discussed along with the operational advantages of dual purpose storage-transport casks. Dual purpose casks can be used for storage at utility and DOE sites (MRS or repository) and for shipment between these sites with minimal spent fuel handling. The cost for a standardized system of casks that are compatible for use at both DOE and utility sites is discussed along with possible arrangements for sharing both the cost and benefits of dual purpose storage-transport casks

  8. ESO Catalogue Facility Design and Performance

    Science.gov (United States)

    Moins, C.; Retzlaff, J.; Arnaboldi, M.; Zampieri, S.; Delmotte, N.; Forchí, V.; Klein Gebbinck, M.; Lockhart, J.; Micol, A.; Vera Sequeiros, I.; Bierwirth, T.; Peron, M.; Romaniello, M.; Suchar, D.

    2013-10-01

    The ESO Phase 3 Catalogue Facility provides investigators with the possibility to ingest catalogues resulting from ESO public surveys and large programs and to query and download their content according to positional and non-positional criteria. It relies on a chain of tools that covers the complete workflow from submission to validation and ingestion into the ESO archive and catalogue repository and a web application to browse and query catalogues. This repository consists of two components. One is a Sybase ASE relational database where catalogue meta-data are stored. The second one is a Sybase IQ data warehouse where the content of each catalogue is ingested in a specific table that returns all records matching a user's query. Spatial indexing has been implemented in Sybase IQ to speed up positional queries and relies on the Spherical Geometry Toolkit from the Johns Hopkins University which implements the Hierarchical Triangular Mesh (HTM) algorithm. It is based on a recursive decomposition of the celestial sphere in spherical triangles and the assignment of an index to each of them. It has been complemented with the use of optimized indexes on the non-positional columns that are likely to be frequently used as query constraints. First tests performed on catalogues such as 2MASS have confirmed that this approach provides a very good level of performance and a smooth user experience that are likely to facilitate the scientific exploitation of catalogues.

  9. NCI's national environmental research data collection: metadata management built on standards and preparing for the semantic web

    Science.gov (United States)

    Wang, Jingbo; Bastrakova, Irina; Evans, Ben; Gohar, Kashif; Santana, Fabiana; Wyborn, Lesley

    2015-04-01

    National Computational Infrastructure (NCI) manages national environmental research data collections (10+ PB) as part of its specialized high performance data node of the Research Data Storage Infrastructure (RDSI) program. We manage 40+ data collections using NCI's Data Management Plan (DMP), which is compatible with the ISO 19100 metadata standards. We utilize ISO standards to make sure our metadata is transferable and interoperable for sharing and harvesting. The DMP is used along with metadata from the data itself, to create a hierarchy of data collection, dataset and time series catalogues that is then exposed through GeoNetwork for standard discoverability. This hierarchy catalogues are linked using a parent-child relationship. The hierarchical infrastructure of our GeoNetwork catalogues system aims to address both discoverability and in-house administrative use-cases. At NCI, we are currently improving the metadata interoperability in our catalogue by linking with standardized community vocabulary services. These emerging vocabulary services are being established to help harmonise data from different national and international scientific communities. One such vocabulary service is currently being established by the Australian National Data Services (ANDS). Data citation is another important aspect of the NCI data infrastructure, which allows tracking of data usage and infrastructure investment, encourage data sharing, and increasing trust in research that is reliant on these data collections. We incorporate the standard vocabularies into the data citation metadata so that the data citation become machine readable and semantically friendly for web-search purpose as well. By standardizing our metadata structure across our entire data corpus, we are laying the foundation to enable the application of appropriate semantic mechanisms to enhance discovery and analysis of NCI's national environmental research data information. We expect that this will further

  10. Improvement Of Search Process In Electronic Catalogues

    Directory of Open Access Journals (Sweden)

    Titas Savickas

    2014-05-01

    Full Text Available The paper presents investigation on search in electronic catalogues. The chosen problem domain is the search system in the electronic catalogue of Lithuanian Academic Libraries. The catalogue uses ALEPH system with MARC21 bibliographic format. The article presents analysis of problems pertaining to the current search engine and user expectations related to the search system of the electronic catalogue of academic libraries. Subsequent to analysis, the research paper presents the architecture for a semantic search system in the electronic catalogue that uses search process designed to improve search results for users.

  11. Examining the Effect of Enactment of a Geospatial Curriculum on Students' Geospatial Thinking and Reasoning

    Science.gov (United States)

    Bodzin, Alec M.; Fu, Qiong; Kulo, Violet; Peffer, Tamara

    2014-08-01

    A potential method for teaching geospatial thinking and reasoning (GTR) is through geospatially enabled learning technologies. We developed an energy resources geospatial curriculum that included learning activities with geographic information systems and virtual globes. This study investigated how 13 urban middle school teachers implemented and varied the enactment of the curriculum with their students and investigated which teacher- and student-level factors accounted for students' GTR posttest achievement. Data included biweekly implementation surveys from teachers and energy resources content and GTR pre- and posttest achievement measures from 1,049 students. Students significantly increased both their energy resources content knowledge and their GTR skills related to energy resources at the end of the curriculum enactment. Both multiple regression and hierarchical linear modeling found that students' initial GTR abilities and gain in energy content knowledge were significantly explanatory variables for their geospatial achievement at the end of curriculum enactment, p critical components of the curriculum or the number of years the teachers had taught the curriculum, did not have significant effects on students' geospatial posttest achievement. The findings from this study provide support that learning with geospatially enabled learning technologies can support GTR with urban middle-level learners.

  12. Technology Catalogue

    International Nuclear Information System (INIS)

    1994-02-01

    The Department of Energy's Office of Environmental Restoration and Waste Management (EM) is responsible for remediating its contaminated sites and managing its waste inventory in a safe and efficient manner. EM's Office of Technology Development (OTD) supports applied research and demonstration efforts to develop and transfer innovative, cost-effective technologies to its site clean-up and waste management programs within EM's Office of Environmental Restoration and Office of Waste Management. The purpose of the Technology Catalogue is to provide performance data on OTD-developed technologies to scientists and engineers assessing and recommending technical solutions within the Department's clean-up and waste management programs, as well as to industry, other federal and state agencies, and the academic community. OTD's applied research and demonstration activities are conducted in programs referred to as Integrated Demonstrations (IDs) and Integrated Programs (IPs). The IDs test and evaluate.systems, consisting of coupled technologies, at specific sites to address generic problems, such as the sensing, treatment, and disposal of buried waste containers. The IPs support applied research activities in specific applications areas, such as in situ remediation, efficient separations processes, and site characterization. The Technology Catalogue is a means for communicating the status. of the development of these innovative technologies. The FY93 Technology Catalogue features technologies successfully demonstrated in the field through IDs and sufficiently mature to be used in the near-term. Technologies from the following IDs are featured in the FY93 Technology Catalogue: Buried Waste ID (Idaho National Engineering Laboratory, Idaho); Mixed Waste Landfill ID (Sandia National Laboratories, New Mexico); Underground Storage Tank ID (Hanford, Washington); Volatile organic compound (VOC) Arid ID (Richland, Washington); and VOC Non-Arid ID (Savannah River Site, South Carolina)

  13. On compact galaxies in the UGC catalogue

    International Nuclear Information System (INIS)

    Kogoshvili, N.G.

    1980-01-01

    A problem of separation of compact galaxies in the UGC Catalogue is considered. Value of surface brightness equal to or less than 21sup(m) was used as compactness criterion from a square second of arc. 96 galaxies, which are brighter than 14sup(m)5 satisfy this criterion. Among compact galaxies discovered in the UGC Catalogue 7% are the Zwicky galaxies, 15% belong to the Markarian galaxies and 27% of galaxies are part of a galaxy list with high surface brightness. Considerable divergence in estimates of total share of compact galaxies in the B.A. Worontsov-Veljaminov Morphological Catalogue of Galaxies (MCG) and the UGC Catalogue is noted. This divergence results from systematical underestimation of visible sizes of compact galaxies in the MCG Catalogue as compared with the UGC Catalogue [ru

  14. Geospatial Database for Strata Objects Based on Land Administration Domain Model (ladm)

    Science.gov (United States)

    Nasorudin, N. N.; Hassan, M. I.; Zulkifli, N. A.; Rahman, A. Abdul

    2016-09-01

    Recently in our country, the construction of buildings become more complex and it seems that strata objects database becomes more important in registering the real world as people now own and use multilevel of spaces. Furthermore, strata title was increasingly important and need to be well-managed. LADM is a standard model for land administration and it allows integrated 2D and 3D representation of spatial units. LADM also known as ISO 19152. The aim of this paper is to develop a strata objects database using LADM. This paper discusses the current 2D geospatial database and needs for 3D geospatial database in future. This paper also attempts to develop a strata objects database using a standard data model (LADM) and to analyze the developed strata objects database using LADM data model. The current cadastre system in Malaysia includes the strata title is discussed in this paper. The problems in the 2D geospatial database were listed and the needs for 3D geospatial database in future also is discussed. The processes to design a strata objects database are conceptual, logical and physical database design. The strata objects database will allow us to find the information on both non-spatial and spatial strata title information thus shows the location of the strata unit. This development of strata objects database may help to handle the strata title and information.

  15. The Role of Discrete Global Grid Systems in the Global Statistical Geospatial Framework

    Science.gov (United States)

    Purss, M. B. J.; Peterson, P.; Minchin, S. A.; Bermudez, L. E.

    2016-12-01

    The United Nations Committee of Experts on Global Geospatial Information Management (UN-GGIM) has proposed the development of a Global Statistical Geospatial Framework (GSGF) as a mechanism for the establishment of common analytical systems that enable the integration of statistical and geospatial information. Conventional coordinate reference systems address the globe with a continuous field of points suitable for repeatable navigation and analytical geometry. While this continuous field is represented on a computer in a digitized and discrete fashion by tuples of fixed-precision floating point values, it is a non-trivial exercise to relate point observations spatially referenced in this way to areal coverages on the surface of the Earth. The GSGF states the need to move to gridded data delivery and the importance of using common geographies and geocoding. The challenges associated with meeting these goals are not new and there has been a significant effort within the geospatial community to develop nested gridding standards to tackle these issues over many years. These efforts have recently culminated in the development of a Discrete Global Grid Systems (DGGS) standard which has been developed under the auspices of Open Geospatial Consortium (OGC). DGGS provide a fixed areal based geospatial reference frame for the persistent location of measured Earth observations, feature interpretations, and modelled predictions. DGGS address the entire planet by partitioning it into a discrete hierarchical tessellation of progressively finer resolution cells, which are referenced by a unique index that facilitates rapid computation, query and analysis. The geometry and location of the cell is the principle aspect of a DGGS. Data integration, decomposition, and aggregation is optimised in the DGGS hierarchical structure and can be exploited for efficient multi-source data processing, storage, discovery, transmission, visualization, computation, analysis, and modelling. During

  16. Geospatial Technologies and Geography Education in a Changing World : Geospatial Practices and Lessons Learned

    NARCIS (Netherlands)

    2015-01-01

    Book published by IGU Commission on Geographical Education. It focuses particularly on what has been learned from geospatial projects and research from the past decades of implementing geospatial technologies in formal and informal education.

  17. Enhancing the online discovery of geospatial data through ...

    African Journals Online (AJOL)

    However, geoportals are often known to geoinformation communities only and present technological limitations which make it difficult for general purpose web search engines to discover and index the data catalogued in (or registered with) a geoportal. The mismatch between standard spatial metadata content and the ...

  18. A Geospatial Online Instruction Model

    OpenAIRE

    Athena OWEN-NAGEL; John C. RODGERS III; Shrinidhi AMBINAKUDIGE

    2012-01-01

    The objective of this study is to present a pedagogical model for teaching geospatial courses through an online format and to critique the model’s effectiveness. Offering geospatial courses through an online format provides avenues to a wider student population, many of whom are not able to take traditional on-campus courses. Yet internet-based teaching effectiveness has not yet been clearly demonstrated for geospatial courses. The pedagogical model implemented in this study heavily utilizes ...

  19. The Compatibility of Developed Mathematics Textbooks' Content in Saudi Arabia (Grades 6-8) with NCTM Standards

    Science.gov (United States)

    Alshehri, Mohammed Ali; Ali, Hassan Shawki

    2016-01-01

    This study aimed to investigate the compatibility of developed mathematics textbooks' content (grades 6-8) in Saudi Arabia with NCTM standards in the areas of: number and operations, algebra, geometry, measurement, data analysis and probability. To achieve that goal, a list of (NCTM) standards for grades (6-8) were translated to Arabic language,…

  20. KINGDOM OF SAUDI ARABIA GEOSPATIAL INFORMATION INFRASTRUCTURE – AN INITIAL STUDY

    Directory of Open Access Journals (Sweden)

    S. H. Alsultan

    2015-10-01

    Full Text Available This paper reviews the current Geographic Information System (Longley et al. implementation and status in the Kingdom of Saudi Arabia (KSA. Based on the review, several problems were identified and discussed. The characteristic of these problems show that the country needs a national geospatial centre. As a new initiative for a national geospatial centre, a study is being conducted especially on best practice from other countries, availability of national committee for standards and policies on data sharing, and the best proposed organization structure inside the administration for the KSA. The study also covers the degree of readiness and awareness among the main GIS stakeholders within the country as well as private parties. At the end of this paper, strategic steps for the national geospatial management centre were proposed as the initial output of the study.

  1. Tendencies in the application of the concept of catalogue marketing in Republic of Serbia and the world

    Directory of Open Access Journals (Sweden)

    Zelić Darko

    2010-01-01

    Full Text Available Catalogue marketing is one of the direct marketing channels. This concept implies making a lot of strategic and tactical decisions that determine catalogue's market success. Catalogue sales is most developed in USA (where it originated and in Western Europe. In Serbia, catalogue marketing is applied just in last few years, since big foreign catalog companies started their business in this region. Here, catalogue marketing is at a lower level of development than in the developed countries, and it comprises a minor part of total trade turnover. The positive thing is that now there are laws that regulates this area, which is encouraging for its development. More and more companies in Serbia are presenting and selling its product range through Internet catalogs. The survey, whose results are briefly presented in this article, showed that consumers in Serbia shop less by print catalogues than consumers in developed countries, and that the partition of those who buy through e-catalogues is increasing. With the increase in standard of living, and overcoming the crisis, there is a chance for catalogue marketing to become much more important concept among consumers and companies in Serbia.

  2. Nansat: a Scientist-Orientated Python Package for Geospatial Data Processing

    Directory of Open Access Journals (Sweden)

    Anton A. Korosov

    2016-10-01

    Full Text Available Nansat is a Python toolbox for analysing and processing 2-dimensional geospatial data, such as satellite imagery, output from numerical models, and gridded in-situ data. It is created with strong focus on facilitating research, and development of algorithms and autonomous processing systems. Nansat extends the widely used Geospatial Abstraction Data Library (GDAL by adding scientific meaning to the datasets through metadata, and by adding common functionality for data analysis and handling (e.g., exporting to various data formats. Nansat uses metadata vocabularies that follow international metadata standards, in particular the Climate and Forecast (CF conventions, and the NASA Directory Interchange Format (DIF and Global Change Master Directory (GCMD keywords. Functionality that is commonly needed in scientific work, such as seamless access to local or remote geospatial data in various file formats, collocation of datasets from different sources and geometries, and visualization, is also built into Nansat. The paper presents Nansat workflows, its functional structure, and examples of typical applications.

  3. NREL Information Resources Catalogue 1999

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2000-04-03

    This is the sixth annual catalogue listing documents produced by NREL during the last fiscal year. Each year the catalogue is mailed to state energy offices, DOE support offices, and to anyone looking to find out more information about NREL's activities and publications.

  4. Extending the ISC-GEM Global Earthquake Instrumental Catalogue

    Science.gov (United States)

    Di Giacomo, Domenico; Engdhal, Bob; Storchak, Dmitry; Villaseñor, Antonio; Harris, James

    2015-04-01

    After a 27-month project funded by the GEM Foundation (www.globalquakemodel.org), in January 2013 we released the ISC-GEM Global Instrumental Earthquake Catalogue (1900 2009) (www.isc.ac.uk/iscgem/index.php) as a special product to use for seismic hazard studies. The new catalogue was necessary as improved seismic hazard studies necessitate that earthquake catalogues are homogeneous (to the largest extent possible) over time in their fundamental parameters, such as location and magnitude. Due to time and resource limitation, the ISC-GEM catalogue (1900-2009) included earthquakes selected according to the following time-variable cut-off magnitudes: Ms=7.5 for earthquakes occurring before 1918; Ms=6.25 between 1918 and 1963; and Ms=5.5 from 1964 onwards. Because of the importance of having a reliable seismic input for seismic hazard studies, funding from GEM and two commercial companies in the US and UK allowed us to start working on the extension of the ISC-GEM catalogue both for earthquakes that occurred beyond 2009 and for earthquakes listed in the International Seismological Summary (ISS) which fell below the cut-off magnitude of 6.25. This extension is part of a four-year program that aims at including in the ISC-GEM catalogue large global earthquakes that occurred before the beginning of the ISC Bulletin in 1964. In this contribution we present the updated ISC GEM catalogue, which will include over 1000 more earthquakes that occurred in 2010 2011 and several hundreds more between 1950 and 1959. The catalogue extension between 1935 and 1949 is currently underway. The extension of the ISC-GEM catalogue will also be helpful for regional cross border seismic hazard studies as the ISC-GEM catalogue should be used as basis for cross-checking the consistency in location and magnitude of those earthquakes listed both in the ISC GEM global catalogue and regional catalogues.

  5. BKE: The catalogue of Bunker-Est Vesuvian station

    International Nuclear Information System (INIS)

    Sarao, A.; Peresan, A.; Vaccari, F.; De Natale, G.; Mariano, A.

    2002-06-01

    A catalogue of 9003 earthquakes associated to the Vesuvian volcano activity as recorded at Bunker Est station (BKE), located on Mt. Vesuvius and operated by the Osservatorio Vesuviano (OV), is presented here. The aim of this catalogue is to integrate the information collected in the catalogue compiled by OV that contains the volcanic earthquakes recorded at station OVO (OVO catalogue) since February 1972. A brief statistical description of the data included in the catalogue BKE and of the empirical relations used for the estimate of the magnitude MBKE from the duration is provided, together with the essential information about the catalogue source. A complete list of the events reported in the catalogue BKE and a description of its format is given in the Appendix. The catalogue BKE has been realized with the cooperation of the University of Trieste - Department of Earth Sciences, the International Centre for Theoretical Physics (ICTP) - Structure and Non-linear Dynamics of the earth (SAND) Group, and the Osservatorio Vesuviano of Naples in the framework of the project 'Eruptive Scenarios from Physical Modeling and Experimental Volcanology' funded by INGV. (author)

  6. Geospatial Information Response Team

    Science.gov (United States)

    Witt, Emitt C.

    2010-01-01

    Extreme emergency events of national significance that include manmade and natural disasters seem to have become more frequent during the past two decades. The Nation is becoming more resilient to these emergencies through better preparedness, reduced duplication, and establishing better communications so every response and recovery effort saves lives and mitigates the long-term social and economic impacts on the Nation. The National Response Framework (NRF) (http://www.fema.gov/NRF) was developed to provide the guiding principles that enable all response partners to prepare for and provide a unified national response to disasters and emergencies. The NRF provides five key principles for better preparation, coordination, and response: 1) engaged partnerships, 2) a tiered response, 3) scalable, flexible, and adaptable operations, 4) unity of effort, and 5) readiness to act. The NRF also describes how communities, tribes, States, Federal Government, privatesector, and non-governmental partners apply these principles for a coordinated, effective national response. The U.S. Geological Survey (USGS) has adopted the NRF doctrine by establishing several earth-sciences, discipline-level teams to ensure that USGS science, data, and individual expertise are readily available during emergencies. The Geospatial Information Response Team (GIRT) is one of these teams. The USGS established the GIRT to facilitate the effective collection, storage, and dissemination of geospatial data information and products during an emergency. The GIRT ensures that timely geospatial data are available for use by emergency responders, land and resource managers, and for scientific analysis. In an emergency and response capacity, the GIRT is responsible for establishing procedures for geospatial data acquisition, processing, and archiving; discovery, access, and delivery of data; anticipating geospatial needs; and providing coordinated products and services utilizing the USGS' exceptional pool of

  7. UCI2001: The updated catalogue of Italy

    International Nuclear Information System (INIS)

    Peresan, A.; Panza, G.F.

    2002-05-01

    A new updated earthquake catalogue for the Italian territory, named UCI2001, is described here; it consists of an updated and revised version of the CCI1996 catalogue (Peresan et al., 1997). The revision essentially corresponds to the incorporation of data from the NEIC (National Earthquake Information Centre) and ALPOR (Catalogo delle Alpi Orientali) catalogues, while the updating is performed using the NEIC Preliminary Determinations of Epicenters since 1986. A brief overview of the catalogues used for the monitoring of seismicity in the Italian area is provided, together with the essential information about the structure of the UCI2001 catalogue and a description of its format. A complete list of the events, as on May 1 2002, is given in the Appendix. (author)

  8. Impact of magnitude uncertainties on seismic catalogue properties

    Science.gov (United States)

    Leptokaropoulos, K. M.; Adamaki, A. K.; Roberts, R. G.; Gkarlaouni, C. G.; Paradisopoulou, P. M.

    2018-05-01

    Catalogue-based studies are of central importance in seismological research, to investigate the temporal, spatial and size distribution of earthquakes in specified study areas. Methods for estimating the fundamental catalogue parameters like the Gutenberg-Richter (G-R) b-value and the completeness magnitude (Mc) are well established and routinely applied. However, the magnitudes reported in seismicity catalogues contain measurement uncertainties which may significantly distort the estimation of the derived parameters. In this study, we use numerical simulations of synthetic data sets to assess the reliability of different methods for determining b-value and Mc, assuming the G-R law validity. After contaminating the synthetic catalogues with Gaussian noise (with selected standard deviations), the analysis is performed for numerous data sets of different sample size (N). The noise introduced to the data generally leads to a systematic overestimation of magnitudes close to and above Mc. This fact causes an increase of the average number of events above Mc, which in turn leads to an apparent decrease of the b-value. This may result to a significant overestimation of seismicity rate even well above the actual completeness level. The b-value can in general be reliably estimated even for relatively small data sets (N < 1000) when only magnitudes higher than the actual completeness level are used. Nevertheless, a correction of the total number of events belonging in each magnitude class (i.e. 0.1 unit) should be considered, to deal with the magnitude uncertainty effect. Because magnitude uncertainties (here with the form of Gaussian noise) are inevitable in all instrumental catalogues, this finding is fundamental for seismicity rate and seismic hazard assessment analyses. Also important is that for some data analyses significant bias cannot necessarily be avoided by choosing a high Mc value for analysis. In such cases, there may be a risk of severe miscalculation of

  9. Spatial data content standards for Africa

    CSIR Research Space (South Africa)

    Cooper, Antony K

    2005-11-01

    Full Text Available , they selected 14 standards containing data dictionaries or feature catalogues, and compared their feature types. They have also provided some advice and recommendations on data content standards (particularly for data dictionaries and feature catalogues...

  10. Evidence Based Cataloguing: Moving Beyond the Rules

    Directory of Open Access Journals (Sweden)

    Kathy Carter

    2010-12-01

    Full Text Available Cataloguing is sometimes regarded as a rule-bound, production-based activity that offers little scope for professional judgement and decision-making. In reality, cataloguing involves challenging decisions that can have significant service and financial impacts. The current environment for cataloguing is a maelstrom of changing demands and competing visions for the future. With information-seekers turning en masse to Google and their behaviour receiving greater attention, library vendors are offering “discovery layer” products to replace traditional OPACs, and cataloguers are examining and debating a transformed version of their descriptive cataloguing rules (Resource Description and Access or RDA. In his “Perceptions of the future of cataloging: Is the sky really falling?” (2009, Ivey provides a good summary of this environment. At the same time, myriad new metadata formats and schema are being developed and applied for digital collections in libraries and other institutions. In today’s libraries, cataloguing is no longer limited to management of traditional AACR and MARC-based metadata for traditional library collections. And like their parent institutions, libraries cannot ignore growing pressures to demonstrate accountability and tangible value provided by their services. More than ever, research and an evidence based approach can help guide cataloguing decision-making.

  11. A geospatial search engine for discovering multi-format geospatial data across the web

    Science.gov (United States)

    Christopher Bone; Alan Ager; Ken Bunzel; Lauren Tierney

    2014-01-01

    The volume of publically available geospatial data on the web is rapidly increasing due to advances in server-based technologies and the ease at which data can now be created. However, challenges remain with connecting individuals searching for geospatial data with servers and websites where such data exist. The objective of this paper is to present a publically...

  12. Shared Geospatial Metadata Repository for Ontario University Libraries: Collaborative Approaches

    Science.gov (United States)

    Forward, Erin; Leahey, Amber; Trimble, Leanne

    2015-01-01

    Successfully providing access to special collections of digital geospatial data in academic libraries relies upon complete and accurate metadata. Creating and maintaining metadata using specialized standards is a formidable challenge for libraries. The Ontario Council of University Libraries' Scholars GeoPortal project, which created a shared…

  13. The Geospatial Web and Local Geographical Education

    Science.gov (United States)

    Harris, Trevor M.; Rouse, L. Jesse; Bergeron, Susan J.

    2010-01-01

    Recent innovations in the Geospatial Web represent a paradigm shift in Web mapping by enabling educators to explore geography in the classroom by dynamically using a rapidly growing suite of impressive online geospatial tools. Coupled with access to spatial data repositories and User-Generated Content, the Geospatial Web provides a powerful…

  14. Comparison of a commercial blood cross-matching kit to the standard laboratory method for establishing blood transfusion compatibility in dogs.

    Science.gov (United States)

    Guzman, Leo Roa; Streeter, Elizabeth; Malandra, Allison

    2016-01-01

    To evaluate the accuracy of a commercial blood transfusion cross-match kit when compared to the standard laboratory method for establishing blood transfusion compatibility. A prospective observational in intro study performed from July 2009 to July 2013. Private referral veterinary center. Ten healthy dogs, 11 anemic dogs, and 24 previously transfused dogs. None. Forty-five dogs were enrolled in a prospective study in order to compare the standard blood transfusion cross-match technique to a commercial blood transfusion cross-matching kit. These dogs were divided into 3 different groups that included 10 healthy dogs (control group), 11 anemic dogs in need of a blood transfusion, and 24 sick dogs that were previously transfused. Thirty-five dogs diagnosed with anemia secondary to multiple disease processes were cross-matched using both techniques. All dogs cross-matched via the kit had a compatible major and minor result, whereas 16 dogs out of 45 (35%) had an incompatible cross-match result when the standard laboratory technique was performed. The average time to perform the commercial kit was 15 minutes and this was 3 times shorter than the manual cross-match laboratory technique that averaged 45-50 minutes to complete. While the gel-based cross-match kit is quicker and less technically demanding than standard laboratory cross-match procedures, microagglutination and low-grade hemolysis are difficult to identify by using the gel-based kits. This could result in transfusion reactions if the gel-based kits are used as the sole determinant of blood compatibility prior to transfusion. Based on our results, the standard manual cross-match technique remains the gold standard test to determine blood transfusion compatibility. © Veterinary Emergency and Critical Care Society 2016.

  15. The Open Geospatial Consortium PUCK Standard: Building Sensor Networks with Self-Describing Instruments

    Science.gov (United States)

    O'Reilly, T. C.; Broering, A.; del Rio, J.; Headley, K. L.; Toma, D.; Bermudez, L. E.; Edgington, D.; Fredericks, J.; Manuel, A.

    2012-12-01

    Sensor technology is rapidly advancing, enabling smaller and cheaper instruments to monitor Earth's environment. It is expected that many more kinds and quantities of networked environmental sensors will be deployed in coming years. Knowledge of each instrument's command protocol is required to operate and acquire data from the network. Making sense of these data streams to create an integrated picture of environmental conditions requires that each instrument's data and metadata be accurately processed and that "suspect" data be flagged. Use of standards to operate an instrument and retrieve and describe its data generally simplifies instrument software development, integration, operation and data processing. The Open Geospatial Consortium (OGC) PUCK protocol enables instruments that describe themselves in a standard way. OGC PUCK defines a small "data sheet" that describes key instrument characteristics, and a standard protocol to retrieve the data sheet from the device itself. Data sheet fields include a universal serial number that is unique across all PUCK-compliant instruments. Other fields identify the instrument manufacturer and model. In addition to the data sheet, the instrument may also provide a "PUCK payload" which can contain additional descriptive information (e.g. a SensorML document or IEEE 1451 TEDS), as well as actual instrument "driver" code. Computers on the sensor network can use PUCK protocol to retrieve this information from installed instruments and utilize it appropriately, e.g. to automatically identify, configure and operate the instruments, and acquire and process their data. The protocol is defined for instruments with an RS232 or Ethernet interface. OGC members recently voted to adopt PUCK as a component of the OGC's Sensor Web Enablement (SWE) standards. The protocol is also supported by a consortium of hydrographic instrument manufacturers and has been implemented by several of them (https://sites.google.com/site/soscsite/). Thus far

  16. Geospatial Technology in Geography Education

    NARCIS (Netherlands)

    Muniz Solari, Osvaldo; Demirci, A.; van der Schee, J.A.

    2015-01-01

    The book is presented as an important starting point for new research in Geography Education (GE) related to the use and application of geospatial technologies (GSTs). For this purpose, the selection of topics was based on central ideas to GE in its relationship with GSTs. The process of geospatial

  17. BioCatalogue: a universal catalogue of web services for the life sciences.

    Science.gov (United States)

    Bhagat, Jiten; Tanoh, Franck; Nzuobontane, Eric; Laurent, Thomas; Orlowski, Jerzy; Roos, Marco; Wolstencroft, Katy; Aleksejevs, Sergejs; Stevens, Robert; Pettifer, Steve; Lopez, Rodrigo; Goble, Carole A

    2010-07-01

    The use of Web Services to enable programmatic access to on-line bioinformatics is becoming increasingly important in the Life Sciences. However, their number, distribution and the variable quality of their documentation can make their discovery and subsequent use difficult. A Web Services registry with information on available services will help to bring together service providers and their users. The BioCatalogue (http://www.biocatalogue.org/) provides a common interface for registering, browsing and annotating Web Services to the Life Science community. Services in the BioCatalogue can be described and searched in multiple ways based upon their technical types, bioinformatics categories, user tags, service providers or data inputs and outputs. They are also subject to constant monitoring, allowing the identification of service problems and changes and the filtering-out of unavailable or unreliable resources. The system is accessible via a human-readable 'Web 2.0'-style interface and a programmatic Web Service interface. The BioCatalogue follows a community approach in which all services can be registered, browsed and incrementally documented with annotations by any member of the scientific community.

  18. A catalogue of crude oil and oil product properties, 1990

    International Nuclear Information System (INIS)

    Bobra, M.A.; Callaghan, S.

    1990-09-01

    This catalogue is a compilation of available data on crude oils and petroleum products. The emphasis of the catalogue is upon oils which could potentially impact Canada's environment. Other oils which are unlikely to be of direct Canadian concern are also included because they have been well characterized and used in relevant studies. The properties listed for each oil are those which will provide an indication of a spilled oil's environmental behaviour and effects. The properties on which data is provided include API gravity, density, viscosity, interfacial tension, pour point, flash point, vapor pressure, volatility and component distribution, emulsion formation tendency and stability, weathering, dispersability, major hydrocarbon groups, aqueous solubility, toxicity, sulfur content, fire point, and wax content. Most of the chemical-physical properties listed in this catalogue were measured using standard tests. For certain properties, data are given at different temperatures and for different degrees of oil weathering. An oil's degree of weathering is expresed as the volume or weight percent evaporated from the fresh oil. Weathered oils used for testing were artificially weathered by gas stripping following the method of Mackay and Stiver. 109 refs

  19. A catalogue of crude oil and oil product properties, 1992

    International Nuclear Information System (INIS)

    Whiticar, S.; Bobra, M.; Liuzzo, P.; Callaghan, S.; Fingas, M.; Jokuty, P.; Ackerman, F.; Cao, J.

    1993-02-01

    This catalogue is a compilation of available data on crude oils and petroleum products. The emphasis of the catalogue is upon oils which could potentially impact Canada's environment. Other oils which are unlikely to be of direct Canadian concern are also included because they have been well characterized and used in relevant studies. The properties listed for each oil are those which will provide an indication of a spilled oil's environmental behaviour and effects. The properties on which data is provided include API gravity, density, viscosity, interfacial tension, pour point, flash point, vapor pressure, volatility and component distribution, emulsion formation tendency and stability, weathering, dispersability, major hydrocarbon groups, aqueous solubility, toxicity, sulfur content, fire point, and wax content. Most of the chemical-physical properties listed in this catalogue were measured using standard tests. For certain properties, data are given at different temperatures and for different degrees of oil weathering. An oil's degree of weathering is expresed as the volume or weight percent evaporated from the fresh oil. Weathered oils used for testing were artificially weathered by gas stripping following the method of Mackay and Stiver. 140 refs

  20. An All-Sky Portable (ASP) Optical Catalogue

    Science.gov (United States)

    Flesch, Eric Wim

    2017-06-01

    This optical catalogue combines the all-sky USNO-B1.0/A1.0 and most-sky APM catalogues, plus overlays of SDSS optical data, into a single all-sky map presented in a sparse binary format that is easily downloaded at 9 Gb zipped. Total count is 1 163 237 190 sources and each has J2000 astrometry, red and blue magnitudes with PSFs and variability indicator, and flags for proper motion, epoch, and source survey and catalogue for each of the photometry and astrometry. The catalogue is available on http://quasars.org/asp.html, and additional data for this paper is available at http://dx.doi.org/10.4225/50/5807fbc12595f.

  1. Geospatial Applications on Different Parallel and Distributed Systems in enviroGRIDS Project

    Science.gov (United States)

    Rodila, D.; Bacu, V.; Gorgan, D.

    2012-04-01

    The execution of Earth Science applications and services on parallel and distributed systems has become a necessity especially due to the large amounts of Geospatial data these applications require and the large geographical areas they cover. The parallelization of these applications comes to solve important performance issues and can spread from task parallelism to data parallelism as well. Parallel and distributed architectures such as Grid, Cloud, Multicore, etc. seem to offer the necessary functionalities to solve important problems in the Earth Science domain: storing, distribution, management, processing and security of Geospatial data, execution of complex processing through task and data parallelism, etc. A main goal of the FP7-funded project enviroGRIDS (Black Sea Catchment Observation and Assessment System supporting Sustainable Development) [1] is the development of a Spatial Data Infrastructure targeting this catchment region but also the development of standardized and specialized tools for storing, analyzing, processing and visualizing the Geospatial data concerning this area. For achieving these objectives, the enviroGRIDS deals with the execution of different Earth Science applications, such as hydrological models, Geospatial Web services standardized by the Open Geospatial Consortium (OGC) and others, on parallel and distributed architecture to maximize the obtained performance. This presentation analysis the integration and execution of Geospatial applications on different parallel and distributed architectures and the possibility of choosing among these architectures based on application characteristics and user requirements through a specialized component. Versions of the proposed platform have been used in enviroGRIDS project on different use cases such as: the execution of Geospatial Web services both on Web and Grid infrastructures [2] and the execution of SWAT hydrological models both on Grid and Multicore architectures [3]. The current

  2. Gaia Data Release 1. Catalogue validation

    NARCIS (Netherlands)

    Arenou, F.; Luri, X.; Babusiaux, C.; Fabricius, C.; Helmi, A.; Robin, A. C.; Vallenari, A.; Blanco-Cuaresma, S.; Cantat-Gaudin, T.; Findeisen, K.; Reylé, C.; Ruiz-Dern, L.; Sordo, R.; Turon, C.; Walton, N. A.; Shih, I.-C.; Antiche, E.; Barache, C.; Barros, M.; Breddels, M.; Carrasco, J. M.; Costigan, G.; Diakité, S.; Eyer, L.; Figueras, F.; Galluccio, L.; Heu, J.; Jordi, C.; Krone-Martins, A.; Lallement, R.; Lambert, S.; Leclerc, N.; Marrese, P. M.; Moitinho, A.; Mor, R.; Romero-Gómez, M.; Sartoretti, P.; Soria, S.; Soubiran, C.; Souchay, J.; Veljanoski, J.; Ziaeepour, H.; Giuffrida, G.; Pancino, E.; Bragaglia, A.

    Context. Before the publication of the Gaia Catalogue, the contents of the first data release have undergone multiple dedicated validation tests. Aims: These tests aim to provide in-depth analysis of the Catalogue content in order to detect anomalies and individual problems in specific objects or in

  3. OSGeo - Open Source Geospatial Foundation

    Directory of Open Access Journals (Sweden)

    Margherita Di Leo

    2012-09-01

    Full Text Available L'esigenza nata verso la fine del 2005 di selezionare ed organizzare più di 200 progetti FOSS4G porta alla nascita nel Febbraio2006 di OSGeo (the Open Source Geospatial Foundation, organizzazione internazionale la cui mission è promuovere lo sviluppo collaborativo di software libero focalizzato sull'informazione geografica (FOSS4G.Open   Source   Geospatial   Foundation (OSGeoThe Open Source Geospatial Foundation (OSGeo  is  a  not-for-profit  organization, created  in  early  2006  to  the  aim  at  sup-porting   the   collaborative   development of  geospatial  open  source  software,  and promote its widespread use. The founda-tion provides financial, organizational and legal support to the broader open source geospatial community. It also serves as an independent  legal  entity  to  which  com-munity  members  can  contribute  code, funding  and  other  resources,  secure  in the knowledge that their contributions will be maintained for public benefit. OSGeo also  serves  as  an  outreach  and  advocacy organization for the open source geospa-tial  community,  and  provides  a  common forum  and  shared  infrastructure  for  im-proving  cross-project  collaboration.  The foundation's projects are all freely available and  useable  under  an  OSI-certified  open source license. The Italian OSGeo local chapter is named GFOSS.it     (Associazione     Italiana     per l'informazione Geografica Libera.

  4. A Geospatial Online Instruction Model

    Science.gov (United States)

    Rodgers, John C., III; Owen-Nagel, Athena; Ambinakudige, Shrinidhi

    2012-01-01

    The objective of this study is to present a pedagogical model for teaching geospatial courses through an online format and to critique the model's effectiveness. Offering geospatial courses through an online format provides avenues to a wider student population, many of whom are not able to take traditional on-campus courses. Yet internet-based…

  5. From Geomatics to Geospatial Intelligent Service Science

    Directory of Open Access Journals (Sweden)

    LI Deren

    2017-10-01

    Full Text Available The paper reviews the 60 years of development from traditional surveying and mapping to today's geospatial intelligent service science.The three important stages of surveying and mapping, namely analogue,analytical and digital stage are summarized.The author introduces the integration of GNSS,RS and GIS(3S,which forms the rise of geospatial informatics(Geomatics.The development of geo-spatial information science in digital earth era is analyzed,and the latest progress of geo-spatial information science towards real-time intelligent service in smart earth era is discussed.This paper focuses on the three development levels of "Internet plus" spatial information intelligent service.In the era of big data,the traditional geomatics will surely take advantage of the integration of communication,navigation,remote sensing,artificial intelligence,virtual reality and brain cognition science,and become geospatial intelligent service science,thereby making contributions to national economy,defense and people's livelihood.

  6. Constitutional compatibility of energy systems

    International Nuclear Information System (INIS)

    Rossnagel, A.

    1983-01-01

    The paper starts from the results of the Enquiry Commission on 'Future Nuclear Energy Policy' of the 8th Federal German Parliament outlining technically feasible energy futures in four 'pathways'. For the purpose of the project, which was to establish the comparative advantages and disadvantages of different energy systems, these four scenarios were reduced to two alternatives: cases K (= nuclear energy) and S (= solar energy). The question to Ge put is: Which changes within our legal system will be ushered in by certain technological developments and how do these changes relate to the legal condition intended so far. Proceeding in this manner will not lead to the result of a nuclear energy system or a solar energy system being in conformity or in contradiction with the constitutional law, but will provide a catalogue of implications orientated to the aims of legal standards: a person deciding in favour of a nuclear energy system or a solar energy system supports this or that development of constitutional policy, and a person purishing this or that aim of legal policy should be consistent and decide in favour of this or that energy system. The investigation of constitutional compatibility leads to the question what effects different energy systems will have on the forms of political intercourse laid down in the constitutional law, which are orientated to models of a liberal constitutional tradition of citizens. (orig./HSCH) [de

  7. Data Quality, Provenance and IPR Management services: their role in empowering geospatial data suppliers and users

    Science.gov (United States)

    Millard, Keiran

    2015-04-01

    This paper looks at current experiences of geospatial users and geospatial suppliers and how they have been limited by suitable frameworks for managing and communicating data quality, data provenance and intellectual property rights (IPR). Current political and technological drivers mean that increasing volumes of geospatial data are available through a plethora of different products and services, and whilst this is inherently a good thing it does create a new generation of challenges. This paper consider two examples of where these issues have been examined and looks at the challenges and possible solutions from a data user and data supplier perspective. The first example is the IQmulus project that is researching fusion environments for big geospatial point clouds and coverages. The second example is the EU Emodnet programme that is establishing thematic data portals for public marine and coastal data. IQmulus examines big geospatial data; the data from sources such as LIDAR, SONAR and numerical simulations; these data are simply too big for routine and ad-hoc analysis, yet they could realise a myriad of disparate, and readily useable, information products with the right infrastructure in place. IQmulus is researching how to deliver this infrastructure technically, but a financially sustainable delivery depends on being able to track and manage ownership and IPR across the numerous data sets being processed. This becomes complex when the data is composed of multiple overlapping coverages, however managing this allows for uses to be delivered highly-bespoke products to meet their budget and technical needs. The Emodnet programme delivers harmonised marine data at the EU scale across seven thematic portals. As part of the Emodnet programme a series of 'check points' have been initiated to examine how useful these services and other public data services actually are to solve real-world problems. One key finding is that users have been confused by the fact that often

  8. Geospatial Information Relevant to the Flood Protection Available on The Mainstream Web

    Directory of Open Access Journals (Sweden)

    Kliment Tomáš

    2014-03-01

    Full Text Available Flood protection is one of several disciplines where geospatial data is very important and is a crucial component. Its management, processing and sharing form the foundation for their efficient use; therefore, special attention is required in the development of effective, precise, standardized, and interoperable models for the discovery and publishing of data on the Web. This paper describes the design of a methodology to discover Open Geospatial Consortium (OGC services on the Web and collect descriptive information, i.e., metadata in a geocatalogue. A pilot implementation of the proposed methodology - Geocatalogue of geospatial information provided by OGC services discovered on Google (hereinafter “Geocatalogue” - was used to search for available resources relevant to the area of flood protection. The result is an analysis of the availability of resources discovered through their metadata collected from the OGC services (WMS, WFS, etc. and the resources they provide (WMS layers, WFS objects, etc. within the domain of flood protection.

  9. Library catalogues as resources for book history: case study of Novosel’s bookstore catalogue in Zagreb (1794 - 1825

    Directory of Open Access Journals (Sweden)

    Marijana Tomić

    2008-07-01

    Full Text Available The aim of the paper is to analyze the book catalogue of Novosel’s bookstore, which operated in Zagreb from 1794 to 1825, and investigate the history of books and writing in Zagreb at the turn of the 19th century. The catalogue we analyzed is believed to have been published in 1801. Bearing in mind that the market-based economy started to develop in the late 18th century, it can be stipulated that Novosel and his staff and successors based the offer in their bookstore on market analysis, i.e. their readers’ needs. The increase in offer has sparked off new advertising techniques, i.e. printing of catalogues. It follows that their book catalogue reflects the image of the cultural and intellectual status and needs of readers in those times. The paper provides a short overview of book trade in the late 18th century Zagreb and of bookstore advertisements published both in books and individually, as well as a short overview of Novosel’s bookstore business. In the analysis we partly use the methodology introduced by Robert Darnton, the so-called Darnton’s circle, which takes a holistic view of the history of books taking into consideration all stages a book needs to go through - from the author, publisher, printer, bookstores, to readers, including the author him/herself as a reader. Every element is considered in relation to other elements in the circle, and in connection with external factors such as the economic and social environment, and political and intellectual influences. The books presented in the catalogue have been analyzed using different criteria: language, genre and country where they were printed. Books printed in Croatia and those written in Croatian have been given priority. In the catalogue analysis we used the database Skupni katalog hrvatskih knjižnica (joint Croatian library catalogue in order to reconstruct the printing year and printing shops that have not been listed in the catalogues. Using this methodology, we partly

  10. Cataloguing In Special Libraries In The 1990s

    Directory of Open Access Journals (Sweden)

    Elizabeth Makin

    1996-01-01

    Full Text Available Cataloguing in special libraries has been virtually ignored in the literature since the turn of the century, although there are many books and papers on cataloguing in general. It is not clear why this should be so, since it can be argued that the needs of special libraries are different from those of public, academic and national libraries. Special libraries are primarily interested in the information content of documents in the sense that they have little or no interest in documents except as "packages" in which information may be encapsulated. It is therefore reasonable to assume, a priori, that special libraries would undertake detailed indexing and light cataloguing, perhaps reducing the catalogue to the status of a finding list. This paper reports the results of a survey of current cataloguing practice in special libraries.

  11. A Global Geospatial Database of 5000+ Historic Flood Event Extents

    Science.gov (United States)

    Tellman, B.; Sullivan, J.; Doyle, C.; Kettner, A.; Brakenridge, G. R.; Erickson, T.; Slayback, D. A.

    2017-12-01

    A key dataset that is missing for global flood model validation and understanding historic spatial flood vulnerability is a global historical geo-database of flood event extents. Decades of earth observing satellites and cloud computing now make it possible to not only detect floods in near real time, but to run these water detection algorithms back in time to capture the spatial extent of large numbers of specific events. This talk will show results from the largest global historical flood database developed to date. We use the Dartmouth Flood Observatory flood catalogue to map over 5000 floods (from 1985-2017) using MODIS, Landsat, and Sentinel-1 Satellites. All events are available for public download via the Earth Engine Catalogue and via a website that allows the user to query floods by area or date, assess population exposure trends over time, and download flood extents in geospatial format.In this talk, we will highlight major trends in global flood exposure per continent, land use type, and eco-region. We will also make suggestions how to use this dataset in conjunction with other global sets to i) validate global flood models, ii) assess the potential role of climatic change in flood exposure iii) understand how urbanization and other land change processes may influence spatial flood exposure iv) assess how innovative flood interventions (e.g. wetland restoration) influence flood patterns v) control for event magnitude to assess the role of social vulnerability and damage assessment vi) aid in rapid probabilistic risk assessment to enable microinsurance markets. Authors on this paper are already using the database for the later three applications and will show examples of wetland intervention analysis in Argentina, social vulnerability analysis in the USA, and micro insurance in India.

  12. Updated earthquake catalogue for seismic hazard analysis in Pakistan

    Science.gov (United States)

    Khan, Sarfraz; Waseem, Muhammad; Khan, Muhammad Asif; Ahmed, Waqas

    2018-03-01

    A reliable and homogenized earthquake catalogue is essential for seismic hazard assessment in any area. This article describes the compilation and processing of an updated earthquake catalogue for Pakistan. The earthquake catalogue compiled in this study for the region (quadrangle bounded by the geographical limits 40-83° N and 20-40° E) includes 36,563 earthquake events, which are reported as 4.0-8.3 moment magnitude (M W) and span from 25 AD to 2016. Relationships are developed between the moment magnitude and body, and surface wave magnitude scales to unify the catalogue in terms of magnitude M W. The catalogue includes earthquakes from Pakistan and neighbouring countries to minimize the effects of geopolitical boundaries in seismic hazard assessment studies. Earthquakes reported by local and international agencies as well as individual catalogues are included. The proposed catalogue is further used to obtain magnitude of completeness after removal of dependent events by using four different algorithms. Finally, seismicity parameters of the seismic sources are reported, and recommendations are made for seismic hazard assessment studies in Pakistan.

  13. Catalogue of Korean manuscripts and rare books

    DEFF Research Database (Denmark)

    Lerbæk Pedersen, Bent

    2014-01-01

    Catalogue of Korean manuscripts and rare books in The Royal Library, Copenhagen and the National Museum of Denmark......Catalogue of Korean manuscripts and rare books in The Royal Library, Copenhagen and the National Museum of Denmark...

  14. A mini-catalogue of metal corrosion products studied by Raman microscopy

    International Nuclear Information System (INIS)

    Bouchard, M.; Smith, D.C.

    2000-01-01

    Full text.The extensive development of physical methods of analysis since the beginning of this century has revolutionised the classical observation techniques most frequently used by the archaeologist. Raman Microscopy (RM) appears to be one of the most promising tools due to the many advantages that it offers: e.g. non-destructive, in situ, micro-analysis. RM is being applied to many archaeological fields as well as to industrial or environmental sectors. In relation with parallel studies made on the identification of corrosion products on archaeological materials, and according to the principal condition for the RM characterisation of an unknown product being the comparison of its Raman spectrum with known standard spectra, the essential aim of this study is to build a mini-catalogue of standard corrosion products susceptible to be found on metallic objects; these could be from archaeological as well as from modern contexts. However, it is noted that the identification of a corrosion product may suggest either an urgent intervention from the restoration team (in the case of active corrosion products), or a stabilisation of the corrosion layer if this is considered to be a protective layer. All the standard samples are natural minerals coming from the Museum National d'Histoire Naturelle in Paris (France) and correspond to the corrosion products most frequently found on metals such copper, zinc, lead or tin. These samples have been analyzed by RM and also confirmed by powder x-ray diffraction analysis. This catalogue, including more than 30 standard species corresponding to the most common metal corrosion products, is very useful for the different studies in progress in collaboration with different archaeological metal restoration teams. The near future will probably see a mobile Raman Microprobe (MRM) equipped with many different mini-catalogues on the site of a corroded mettalic bridge, a corroded canalisation or under the sea to rapidly identify the different

  15. Central Asia earthquake catalogue from ancient time to 2009

    Directory of Open Access Journals (Sweden)

    Natalya N. Mikhailova

    2015-04-01

    Full Text Available In this work, we present the seismic catalogue compiled for Central Asia (Kazakhstan, Kyrgyzstan, Tajikistan, Uzbekistan and Turkmenistan in the framework of the Earthquake Model Central Asia (EMCA project. The catalogue from 2000 B.C. to 2009 A.D. is composed by 33,034 earthquakes in the MLH magnitude (magnitude by surface waves on horizontal components widely used in practice of the former USSR countries range from 1.5 to 8.3. The catalogue includes both macroseimic and instrumental constrained data, with about 32,793 earthquake after 1900 A.D. The main sources and procedure used to compile the catalogues are discussed, and the comparison with the ISC-GEM catalogue presented. Magnitude of completeness analysis shows that the catalogue is complete down to magnitude 4 from 1959 and to magnitude 7 from 1873, whereas the obtained regional b value is 0.805.

  16. Geospatial Image Stream Processing: Models, techniques, and applications in remote sensing change detection

    Science.gov (United States)

    Rueda-Velasquez, Carlos Alberto

    Detection of changes in environmental phenomena using remotely sensed data is a major requirement in the Earth sciences, especially in natural disaster related scenarios where real-time detection plays a crucial role in the saving of human lives and the preservation of natural resources. Although various approaches formulated to model multidimensional data can in principle be applied to the inherent complexity of remotely sensed geospatial data, there are still challenging peculiarities that demand a precise characterization in the context of change detection, particularly in scenarios of fast changes. In the same vein, geospatial image streams do not fit appropriately in the standard Data Stream Management System (DSMS) approach because these systems mainly deal with tuple-based streams. Recognizing the necessity for a systematic effort to address the above issues, the work presented in this thesis is a concrete step toward the foundation and construction of an integrated Geospatial Image Stream Processing framework, GISP. First, we present a data and metadata model for remotely sensed image streams. We introduce a precise characterization of images and image streams in the context of remotely sensed geospatial data. On this foundation, we define spatially-aware temporal operators with a consistent semantics for change analysis tasks. We address the change detection problem in settings where multiple image stream sources are available, and thus we introduce an architectural design for the processing of geospatial image streams from multiple sources. With the aim of targeting collaborative scientific environments, we construct a realization of our architecture based on Kepler, a robust and widely used scientific workflow management system, as the underlying computational support; and open data and Web interface standards, as a means to facilitate the interoperability of GISP instances with other processing infrastructures and client applications. We demonstrate our

  17. Visualization and Ontology of Geospatial Intelligence

    Science.gov (United States)

    Chan, Yupo

    Recent events have deepened our conviction that many human endeavors are best described in a geospatial context. This is evidenced in the prevalence of location-based services, as afforded by the ubiquitous cell phone usage. It is also manifested by the popularity of such internet engines as Google Earth. As we commute to work, travel on business or pleasure, we make decisions based on the geospatial information provided by such location-based services. When corporations devise their business plans, they also rely heavily on such geospatial data. By definition, local, state and federal governments provide services according to geographic boundaries. One estimate suggests that 85 percent of data contain spatial attributes.

  18. Catalogue of HI PArameters (CHIPA)

    Science.gov (United States)

    Saponara, J.; Benaglia, P.; Koribalski, B.; Andruchow, I.

    2015-08-01

    The catalogue of HI parameters of galaxies HI (CHIPA) is the natural continuation of the compilation by M.C. Martin in 1998. CHIPA provides the most important parameters of nearby galaxies derived from observations of the neutral Hydrogen line. The catalogue contains information of 1400 galaxies across the sky and different morphological types. Parameters like the optical diameter of the galaxy, the blue magnitude, the distance, morphological type, HI extension are listed among others. Maps of the HI distribution, velocity and velocity dispersion can also be display for some cases. The main objective of this catalogue is to facilitate the bibliographic queries, through searching in a database accessible from the internet that will be available in 2015 (the website is under construction). The database was built using the open source `` mysql (SQL, Structured Query Language, management system relational database) '', while the website was built with ''HTML (Hypertext Markup Language)'' and ''PHP (Hypertext Preprocessor)''.

  19. An approach for heterogeneous and loosely coupled geospatial data distributed computing

    Science.gov (United States)

    Chen, Bin; Huang, Fengru; Fang, Yu; Huang, Zhou; Lin, Hui

    2010-07-01

    Most GIS (Geographic Information System) applications tend to have heterogeneous and autonomous geospatial information resources, and the availability of these local resources is unpredictable and dynamic under a distributed computing environment. In order to make use of these local resources together to solve larger geospatial information processing problems that are related to an overall situation, in this paper, with the support of peer-to-peer computing technologies, we propose a geospatial data distributed computing mechanism that involves loosely coupled geospatial resource directories and a term named as Equivalent Distributed Program of global geospatial queries to solve geospatial distributed computing problems under heterogeneous GIS environments. First, a geospatial query process schema for distributed computing as well as a method for equivalent transformation from a global geospatial query to distributed local queries at SQL (Structured Query Language) level to solve the coordinating problem among heterogeneous resources are presented. Second, peer-to-peer technologies are used to maintain a loosely coupled network environment that consists of autonomous geospatial information resources, thus to achieve decentralized and consistent synchronization among global geospatial resource directories, and to carry out distributed transaction management of local queries. Finally, based on the developed prototype system, example applications of simple and complex geospatial data distributed queries are presented to illustrate the procedure of global geospatial information processing.

  20. Catalogue of Exoplanets in Multiple-Star-Systems

    Science.gov (United States)

    Schwarz, Richard; Funk, Barbara; Bazsó, Ákos; Pilat-Lohinger, Elke

    2017-07-01

    Cataloguing the data of exoplanetary systems becomes more and more important, due to the fact that they conclude the observations and support the theoretical studies. Since 1995 there is a database which list most of the known exoplanets (The Extrasolar Planets Encyclopaedia is available at http://exoplanet.eu/ and described at Schneider et al. 2011). With the growing number of detected exoplanets in binary and multiple star systems it became more important to mark and to separate them into a new database. Therefore we started to compile a catalogue for binary and multiple star systems. Since 2013 the catalogue can be found at http://www.univie.ac.at/adg/schwarz/multiple.html (description can be found at Schwarz et al. 2016) which will be updated regularly and is linked to the Extrasolar Planets Encyclopaedia. The data of the binary catalogue can be downloaded as a file (.csv) and used for statistical purposes. Our database is divided into two parts: the data of the stars and the planets, given in a separate list. Every columns of the list can be sorted in two directions: ascending, meaning from the lowest value to the highest, or descending. In addition an introduction and help is also given in the menu bar of the catalogue including an example list.

  1. Catalogue of Meteor Showers and Storms in Korean History

    Directory of Open Access Journals (Sweden)

    Sang-Hyeon Ahn

    2004-03-01

    Full Text Available We present a more complete and accurate catalogue of astronomical records for meteor showers and meteor storms appeared in primary official Korean history books, such as Samguk-sagi, Koryo-sa, Seungjeongwon-ilgi, and Choson-Wangjo-Sillok. So far the catalogue made by Imoto and Hasegawa in 1958 has been widely used in the international astronomical society. The catalogue is based on a report by Sekiguchi in 1917 that is mainly based on secondary history books. We observed that the catalogue has a number of errors in either dates or sources of the records. We have thoroughly checked the primary official history books, instead of the secondary ones, in order to make a corrected and extended catalogue. The catalogue contains 25 records of meteor storms, four records of intense meteor-showers, and five records of usual showers in Korean history. We also find that some of those records seem to correspond to some presently active meteor showers such as the Leonids, the Perseids, and the ¥ç-Aquarids-Orionids pair. However, a large number of those records do not correspond to such present showers. This catalogue we obtained can be useful for various astrophysical studies in the future.

  2. Building a Disciplinary Metadata Standards Directory

    Directory of Open Access Journals (Sweden)

    Alexander Ball

    2014-07-01

    Full Text Available The Research Data Alliance (RDA Metadata Standards Directory Working Group (MSDWG is building a directory of descriptive, discipline-specific metadata standards. The purpose of the directory is to promote the discovery, access and use of such standards, thereby improving the state of research data interoperability and reducing duplicative standards development work.This work builds upon the UK Digital Curation Centre's Disciplinary Metadata Catalogue, a resource created with much the same aim in mind. The first stage of the MSDWG's work was to update and extend the information contained in the catalogue. In the current, second stage, a new platform is being developed in order to extend the functionality of the directory beyond that of the catalogue, and to make it easier to maintain and sustain. Future work will include making the directory more amenable to use by automated tools.

  3. Describing Geospatial Assets in the Web of Data: A Metadata Management Scenario

    Directory of Open Access Journals (Sweden)

    Cristiano Fugazza

    2016-12-01

    Full Text Available Metadata management is an essential enabling factor for geospatial assets because discovery, retrieval, and actual usage of the latter are tightly bound to the quality of these descriptions. Unfortunately, the multi-faceted landscape of metadata formats, requirements, and conventions makes it difficult to identify editing tools that can be easily tailored to the specificities of a given project, workgroup, and Community of Practice. Our solution is a template-driven metadata editing tool that can be customised to any XML-based schema. Its output is constituted by standards-compliant metadata records that also have a semantics-aware counterpart eliciting novel exploitation techniques. Moreover, external data sources can easily be plugged in to provide autocompletion functionalities on the basis of the data structures made available on the Web of Data. Beside presenting the essentials on customisation of the editor by means of two use cases, we extend the methodology to the whole life cycle of geospatial metadata. We demonstrate the novel capabilities enabled by RDF-based metadata representation with respect to traditional metadata management in the geospatial domain.

  4. Technology Catalogue. First edition

    Energy Technology Data Exchange (ETDEWEB)

    1994-02-01

    The Department of Energy`s Office of Environmental Restoration and Waste Management (EM) is responsible for remediating its contaminated sites and managing its waste inventory in a safe and efficient manner. EM`s Office of Technology Development (OTD) supports applied research and demonstration efforts to develop and transfer innovative, cost-effective technologies to its site clean-up and waste management programs within EM`s Office of Environmental Restoration and Office of Waste Management. The purpose of the Technology Catalogue is to provide performance data on OTD-developed technologies to scientists and engineers assessing and recommending technical solutions within the Department`s clean-up and waste management programs, as well as to industry, other federal and state agencies, and the academic community. OTD`s applied research and demonstration activities are conducted in programs referred to as Integrated Demonstrations (IDs) and Integrated Programs (IPs). The IDs test and evaluate.systems, consisting of coupled technologies, at specific sites to address generic problems, such as the sensing, treatment, and disposal of buried waste containers. The IPs support applied research activities in specific applications areas, such as in situ remediation, efficient separations processes, and site characterization. The Technology Catalogue is a means for communicating the status. of the development of these innovative technologies. The FY93 Technology Catalogue features technologies successfully demonstrated in the field through IDs and sufficiently mature to be used in the near-term. Technologies from the following IDs are featured in the FY93 Technology Catalogue: Buried Waste ID (Idaho National Engineering Laboratory, Idaho); Mixed Waste Landfill ID (Sandia National Laboratories, New Mexico); Underground Storage Tank ID (Hanford, Washington); Volatile organic compound (VOC) Arid ID (Richland, Washington); and VOC Non-Arid ID (Savannah River Site, South Carolina).

  5. Competencies and materials for repositioning cataloguers for ...

    African Journals Online (AJOL)

    The purpose of this study was to determine the competencies and materials for repositioning cataloguers for information management in an electronic era. The survey method was adopted for the research design using questionnaire for data collection. The population comprised of 44 cataloguers in 12 universities in ...

  6. Planck 2015 results. XXVIII. The Planck Catalogue of Galactic Cold Clumps

    CERN Document Server

    Ade, P.A.R.; Arnaud, M.; Ashdown, M.; Aumont, J.; Baccigalupi, C.; Banday, A.J.; Barreiro, R.B.; Bartolo, N.; Battaner, E.; Benabed, K.; Benoît, A.; Benoit-Lévy, A.; Bernard, J.-P.; Bersanelli, M.; Bielewicz, P.; Bonaldi, A.; Bonavera, L.; Bond, J.R.; Borrill, J.; Bouchet, F.R.; Boulanger, F.; Bucher, M.; Burigana, C.; Butler, R.C.; Calabrese, E.; Catalano, A.; Chamballu, A.; Chiang, H.C.; Christensen, P.R.; Clements, D.L.; Colombi, S.; Colombo, L.P.L.; Combet, C.; Couchot, F.; Coulais, A.; Crill, B.P.; Curto, A.; Cuttaia, F.; Danese, L.; Davies, R.D.; Davis, R.J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Désert, F.-X.; Dickinson, C.; Diego, J.M.; Dole, H.; Donzelli, S.; Doré, O.; Douspis, M.; Ducout, A.; Dupac, X.; Efstathiou, G.; Elsner, F.; Enßlin, T.A.; Eriksen, H.K.; Falgarone, E.; Fergusson, J.; Finelli, F.; Forni, O.; Frailis, M.; Fraisse, A.A.; Franceschi, E.; Frejsel, A.; Galeotta, S.; Galli, S.; Ganga, K.; Giard, M.; Giraud-Héraud, Y.; Gjerløw, E.; González-Nuevo, J.; Górski, K.M.; Gratton, S.; Gregorio, A.; Gruppuso, A.; Gudmundsson, J.E.; Hansen, F.K.; Hanson, D.; Harrison, D.L.; Helou, G.; Henrot-Versillé, S.; Hernández-Monteagudo, C.; Herranz, D.; Hildebrandt, S.R.; Hivon, E.; Hobson, M.; Holmes, W.A.; Hornstrup, A.; Hovest, W.; Huffenberger, K.M.; Hurier, G.; Jaffe, A.H.; Jaffe, T.R.; Jones, W.C.; Juvela, M.; Keihänen, E.; Keskitalo, R.; Kisner, T.S.; Knoche, J.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lamarre, J.-M.; Lasenby, A.; Lattanzi, M.; Lawrence, C.R.; Leonardi, R.; Lesgourgues, J.; Levrier, F.; Liguori, M.; Lilje, P.B.; Linden-Vørnle, M.; López-Caniego, M.; Lubin, P.M.; Macías-Pérez, J.F.; Maggio, G.; Maino, D.; Mandolesi, N.; Mangilli, A.; Marshall, D.J.; Martin, P.G.; Martínez-González, E.; Masi, S.; Matarrese, S.; Mazzotta, P.; McGehee, P.; Melchiorri, A.; Mendes, L.; Mennella, A.; Migliaccio, M.; Mitra, S.; Miville-Deschênes, M.-A.; Moneti, A.; Morgante, G.; Mortlock, D.; Moss, A.; Munshi, D.; Murphy, J.A.; Naselsky, P.; Nati, F.; Natoli, P.; Netterfield, C.B.; Nørgaard-Nielsen, H.U.; Noviello, F.; Novikov, D.; Novikov, I.; Oxborrow, C.A.; Paci, F.; Pagano, L.; Pajot, F.; Paladini, R.; Paoletti, D.; Pasian, F.; Patanchon, G.; Pearson, T.J.; Pelkonen, V.-M.; Perdereau, O.; Perotto, L.; Perrotta, F.; Pettorino, V.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Pratt, G.W.; Prézeau, G.; Prunet, S.; Puget, J.-L.; Rachen, J.P.; Reach, W.T.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Renzi, A.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Rossetti, M.; Roudier, G.; Rubiño-Martín, J.A.; Rusholme, B.; Sandri, M.; Santos, D.; Savelainen, M.; Savini, G.; Scott, D.; Seiffert, M.D.; Shellard, E.P.S.; Spencer, L.D.; Stolyarov, V.; Sudiwala, R.; Sunyaev, R.; Sutton, D.; Suur-Uski, A.-S.; Sygnet, J.-F.; Tauber, J.A.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tristram, M.; Tucci, M.; Tuovinen, J.; Umana, G.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Vielva, P.; Villa, F.; Wade, L.A.; Wandelt, B.D.; Wehus, I.K.; Yvon, D.; Zacchei, A.

    2016-01-01

    We present the Planck Catalogue of Galactic Cold Clumps (PGCC), an all-sky catalogue of Galactic cold clump candidates detected by Planck. This catalogue is the full version of the Early Cold Core (ECC) catalogue, which was made available in 2011 with the Early Release Compact Source Catalogue (ERCSC) and contained 915 high S/N sources. It is based on the Planck 48 months mission data that are currently being released to the astronomical community. The PGCC catalogue is an observational catalogue consisting exclusively of Galactic cold sources. The three highest Planck bands (857, 545, 353 GHz) have been combined with IRAS data at 3 THz to perform a multi-frequency detection of sources colder than their local environment. After rejection of possible extragalactic contaminants, the PGCC catalogue contains 13188 Galactic sources spread across the whole sky, i.e., from the Galactic plane to high latitudes, following the spatial distribution of the main molecular cloud complexes. The median temperature of PGCC so...

  7. Economic Assessment of the Use Value of Geospatial Information

    Directory of Open Access Journals (Sweden)

    Richard Bernknopf

    2015-07-01

    Full Text Available Geospatial data inform decision makers. An economic model that involves application of spatial and temporal scientific, technical, and economic data in decision making is described. The value of information (VOI contained in geospatial data is the difference between the net benefits (in present value terms of a decision with and without the information. A range of technologies is used to collect and distribute geospatial data. These technical activities are linked to examples that show how the data can be applied in decision making, which is a cultural activity. The economic model for assessing the VOI in geospatial data for decision making is applied to three examples: (1 a retrospective model about environmental regulation of agrochemicals; (2 a prospective model about the impact and mitigation of earthquakes in urban areas; and (3 a prospective model about developing private–public geospatial information for an ecosystem services market. Each example demonstrates the potential value of geospatial information in a decision with uncertain information.

  8. Economic assessment of the use value of geospatial information

    Science.gov (United States)

    Bernknopf, Richard L.; Shapiro, Carl D.

    2015-01-01

    Geospatial data inform decision makers. An economic model that involves application of spatial and temporal scientific, technical, and economic data in decision making is described. The value of information (VOI) contained in geospatial data is the difference between the net benefits (in present value terms) of a decision with and without the information. A range of technologies is used to collect and distribute geospatial data. These technical activities are linked to examples that show how the data can be applied in decision making, which is a cultural activity. The economic model for assessing the VOI in geospatial data for decision making is applied to three examples: (1) a retrospective model about environmental regulation of agrochemicals; (2) a prospective model about the impact and mitigation of earthquakes in urban areas; and (3) a prospective model about developing private–public geospatial information for an ecosystem services market. Each example demonstrates the potential value of geospatial information in a decision with uncertain information.

  9. Increasing the value of geospatial informatics with open approaches for Big Data

    Science.gov (United States)

    Percivall, G.; Bermudez, L. E.

    2017-12-01

    Open approaches to big data provide geoscientists with new capabilities to address problems of unmatched size and complexity. Consensus approaches for Big Geo Data have been addressed in multiple international workshops and testbeds organized by the Open Geospatial Consortium (OGC) in the past year. Participants came from government (NASA, ESA, USGS, NOAA, DOE); research (ORNL, NCSA, IU, JPL, CRIM, RENCI); industry (ESRI, Digital Globe, IBM, rasdaman); standards (JTC 1/NIST); and open source software communities. Results from the workshops and testbeds are documented in Testbed reports and a White Paper published by the OGC. The White Paper identifies the following set of use cases: Collection and Ingest: Remote sensed data processing; Data stream processing Prepare and Structure: SQL and NoSQL databases; Data linking; Feature identification Analytics and Visualization: Spatial-temporal analytics; Machine Learning; Data Exploration Modeling and Prediction: Integrated environmental models; Urban 4D models. Open implementations were developed in the Arctic Spatial Data Pilot using Discrete Global Grid Systems (DGGS) and in Testbeds using WPS and ESGF to publish climate predictions. Further development activities to advance open implementations of Big Geo Data include the following: Open Cloud Computing: Avoid vendor lock-in through API interoperability and Application portability. Open Source Extensions: Implement geospatial data representations in projects from Apache, Location Tech, and OSGeo. Investigate parallelization strategies for N-Dimensional spatial data. Geospatial Data Representations: Schemas to improve processing and analysis using geospatial concepts: Features, Coverages, DGGS. Use geospatial encodings like NetCDF and GeoPackge. Big Linked Geodata: Use linked data methods scaled to big geodata. Analysis Ready Data: Support "Download as last resort" and "Analytics as a service". Promote elements common to "datacubes."

  10. Second ROSAT all-sky survey (2RXS) source catalogue

    Science.gov (United States)

    Boller, Th.; Freyberg, M. J.; Trümper, J.; Haberl, F.; Voges, W.; Nandra, K.

    2016-04-01

    Aims: We present the second ROSAT all-sky survey source catalogue, hereafter referred to as the 2RXS catalogue. This is the second publicly released ROSAT catalogue of point-like sources obtained from the ROSAT all-sky survey (RASS) observations performed with the position-sensitive proportional counter (PSPC) between June 1990 and August 1991, and is an extended and revised version of the bright and faint source catalogues. Methods: We used the latest version of the RASS processing to produce overlapping X-ray images of 6.4° × 6.4° sky regions. To create a source catalogue, a likelihood-based detection algorithm was applied to these, which accounts for the variable point-spread function (PSF) across the PSPC field of view. Improvements in the background determination compared to 1RXS were also implemented. X-ray control images showing the source and background extraction regions were generated, which were visually inspected. Simulations were performed to assess the spurious source content of the 2RXS catalogue. X-ray spectra and light curves were extracted for the 2RXS sources, with spectral and variability parameters derived from these products. Results: We obtained about 135 000 X-ray detections in the 0.1-2.4 keV energy band down to a likelihood threshold of 6.5, as adopted in the 1RXS faint source catalogue. Our simulations show that the expected spurious content of the catalogue is a strong function of detection likelihood, and the full catalogue is expected to contain about 30% spurious detections. A more conservative likelihood threshold of 9, on the other hand, yields about 71 000 detections with a 5% spurious fraction. We recommend thresholds appropriate to the scientific application. X-ray images and overlaid X-ray contour lines provide an additional user product to evaluate the detections visually, and we performed our own visual inspections to flag uncertain detections. Intra-day variability in the X-ray light curves was quantified based on the

  11. A 'new generation' earthquake catalogue

    Directory of Open Access Journals (Sweden)

    E. Boschi

    2000-06-01

    Full Text Available In 1995, we published the first release of the Catalogo dei Forti Terremoti in Italia, 461 a.C. - 1980, in Italian (Boschi et al., 1995. Two years later this was followed by a second release, again in Italian, that included more earthquakes, more accurate research and a longer time span (461 B.C. to 1990 (Boschi et al., 1997. Aware that the record of Italian historical seismicity is probably the most extensive of the whole world, and hence that our catalogue could be of interest for a wider interna-tional readership, Italian was clearly not the appropriate language to share this experience with colleagues from foreign countries. Three years after publication of the second release therefore, and after much additional research and fine tuning of methodologies and algorithms, I am proud to introduce this third release in English. All the tools and accessories have been translated along with the texts describing the development of the underlying research strategies and current contents. The English title is Catalogue of Strong Italian Earthquakes, 461 B.C. to 1997. This Preface briefly describes the scientific context within which the Catalogue of Strong Italian Earthquakes was conceived and progressively developed. The catalogue is perhaps the most impor-tant outcome of a well-established joint project between the Istituto Nazionale di Geofisica, the leading Italian institute for basic and applied research in seismology and solid earth geophysics, and SGA (Storia Geofisica Ambiente, a private firm specialising in the historical investigation and systematisation of natural phenomena. In her contribution "Method of investigation, typology and taxonomy of the basic data: navigating between seismic effects and historical contexts", Emanuela Guidoboni outlines the general framework of modern historical seismology, its complex relation with instrumental seismology on the one hand and historical research on the other. This presentation also highlights

  12. Updating Geospatial Data from Large Scale Data Sources

    Science.gov (United States)

    Zhao, R.; Chen, J.; Wang, D.; Shang, Y.; Wang, Z.; Li, X.; Ai, T.

    2011-08-01

    In the past decades, many geospatial databases have been established at national, regional and municipal levels over the world. Nowadays, it has been widely recognized that how to update these established geo-spatial database and keep them up to date is most critical for the value of geo-spatial database. So, more and more efforts have been devoted to the continuous updating of these geospatial databases. Currently, there exist two main types of methods for Geo-spatial database updating: directly updating with remote sensing images or field surveying materials, and indirectly updating with other updated data result such as larger scale newly updated data. The former method is the basis because the update data sources in the two methods finally root from field surveying and remote sensing. The later method is often more economical and faster than the former. Therefore, after the larger scale database is updated, the smaller scale database should be updated correspondingly in order to keep the consistency of multi-scale geo-spatial database. In this situation, it is very reasonable to apply map generalization technology into the process of geo-spatial database updating. The latter is recognized as one of most promising methods of geo-spatial database updating, especially in collaborative updating environment in terms of map scale, i.e , different scale database are produced and maintained separately by different level organizations such as in China. This paper is focused on applying digital map generalization into the updating of geo-spatial database from large scale in the collaborative updating environment for SDI. The requirements of the application of map generalization into spatial database updating are analyzed firstly. A brief review on geospatial data updating based digital map generalization is then given. Based on the requirements analysis and review, we analyze the key factors for implementing updating geospatial data from large scale including technical

  13. VIRAC: The VVV Infrared Astrometric Catalogue

    OpenAIRE

    Smith, L. C.; Lucas, P. W.; Kurtev, R.; Smart, R.; Minniti, D.; Borissova, J.; Jones, H. R. A; Zhang, Z. H.; Marocco, F.; Peña, C. Contreras; Gromadzki, M.; Kuhn, M. A.; Drew, J. E.; Pinfield, D. J.; Bedin, L. R.

    2017-01-01

    We present VIRAC version 1, a near-infrared proper motion and parallax catalogue of the VISTA VVV survey for 312,587,642 unique sources averaged across all overlapping pawprint and tile images covering 560 deg$^2$ of the bulge of the Milky Way and southern disk. The catalogue includes 119 million high quality proper motion measurements, of which 47 million have statistical uncertainties below 1 mas yr$^{-1}$. In the 11$

  14. Geospatial Web Services in Real Estate Information System

    Science.gov (United States)

    Radulovic, Aleksandra; Sladic, Dubravka; Govedarica, Miro; Popovic, Dragana; Radovic, Jovana

    2017-12-01

    Since the data of cadastral records are of great importance for the economic development of the country, they must be well structured and organized. Records of real estate on the territory of Serbia met many problems in previous years. To prevent problems and to achieve efficient access, sharing and exchange of cadastral data on the principles of interoperability, domain model for real estate is created according to current standards in the field of spatial data. The resulting profile of the domain model for the Serbian real estate cadastre is based on the current legislation and on Land Administration Domain Model (LADM) which is specified in the ISO19152 standard. Above such organized data, and for their effective exchange, it is necessary to develop a model of services that must be provided by the institutions interested in the exchange of cadastral data. This is achieved by introducing a service-oriented architecture in the information system of real estate cadastre and with that ensures efficiency of the system. It is necessary to develop user services for download, review and use of the real estate data through the web. These services should be provided to all users who need access to cadastral data (natural and legal persons as well as state institutions) through e-government. It is also necessary to provide search, view and download of cadastral spatial data by specifying geospatial services. Considering that real estate contains geometric data for parcels and buildings it is necessary to establish set of geospatial services that would provide information and maps for the analysis of spatial data, and for forming a raster data. Besides the theme Cadastral parcels, INSPIRE directive specifies several themes that involve data on buildings and land use, for which data can be provided from real estate cadastre. In this paper, model of geospatial services in Serbia is defined. A case study of using these services to estimate which household is at risk of

  15. Geospatial Information is the Cornerstone of Effective Hazards Response

    Science.gov (United States)

    Newell, Mark

    2008-01-01

    Every day there are hundreds of natural disasters world-wide. Some are dramatic, whereas others are barely noticeable. A natural disaster is commonly defined as a natural event with catastrophic consequences for living things in the vicinity. Those events include earthquakes, floods, hurricanes, landslides, tsunami, volcanoes, and wildfires. Man-made disasters are events that are caused by man either intentionally or by accident, and that directly or indirectly threaten public health and well-being. These occurrences span the spectrum from terrorist attacks to accidental oil spills. To assist in responding to natural and potential man-made disasters, the U.S. Geological Survey (USGS) has established the Geospatial Information Response Team (GIRT) (http://www.usgs.gov/emergency/). The primary purpose of the GIRT is to ensure rapid coordination and availability of geospatial information for effective response by emergency responders, and land and resource managers, and for scientific analysis. The GIRT is responsible for establishing monitoring procedures for geospatial data acquisition, processing, and archiving; discovery, access, and delivery of data; anticipating geospatial needs; and providing relevant geospatial products and services. The GIRT is focused on supporting programs, offices, other agencies, and the public in mission response to hazards. The GIRT will leverage the USGS Geospatial Liaison Network and partnerships with the Department of Homeland Security (DHS), National Geospatial-Intelligence Agency (NGA), and Northern Command (NORTHCOM) to coordinate the provisioning and deployment of USGS geospatial data, products, services, and equipment. The USGS geospatial liaisons will coordinate geospatial information sharing with State, local, and tribal governments, and ensure geospatial liaison back-up support procedures are in place. The GIRT will coordinate disposition of USGS staff in support of DHS response center activities as requested by DHS. The GIRT

  16. A Python Geospatial Language Toolkit

    Science.gov (United States)

    Fillmore, D.; Pletzer, A.; Galloy, M.

    2012-12-01

    The volume and scope of geospatial data archives, such as collections of satellite remote sensing or climate model products, has been rapidly increasing and will continue to do so in the near future. The recently launched (October 2011) Suomi National Polar-orbiting Partnership satellite (NPP) for instance, is the first of a new generation of Earth observation platforms that will monitor the atmosphere, oceans, and ecosystems, and its suite of instruments will generate several terabytes each day in the form of multi-spectral images and derived datasets. Full exploitation of such data for scientific analysis and decision support applications has become a major computational challenge. Geophysical data exploration and knowledge discovery could benefit, in particular, from intelligent mechanisms for extracting and manipulating subsets of data relevant to the problem of interest. Potential developments include enhanced support for natural language queries and directives to geospatial datasets. The translation of natural language (that is, human spoken or written phrases) into complex but unambiguous objects and actions can be based on a context, or knowledge domain, that represents the underlying geospatial concepts. This poster describes a prototype Python module that maps English phrases onto basic geospatial objects and operations. This module, along with the associated computational geometry methods, enables the resolution of natural language directives that include geographic regions of arbitrary shape and complexity.

  17. Publications catalogue 1982-83

    International Nuclear Information System (INIS)

    1982-04-01

    This catalogue lists the technical reports, papers, speeches, regulatory documents, news releases, information bulletins, notices, and miscellaneous documents issued by the Canadian Atomic Energy Control Board between 1977 and 1982

  18. eGenomics: Cataloguing Our Complete Genome Collection III

    Directory of Open Access Journals (Sweden)

    Dawn Field

    2007-01-01

    Full Text Available This meeting report summarizes the proceedings of the “eGenomics: Cataloguing our Complete Genome Collection III” workshop held September 11–13, 2006, at the National Institute for Environmental eScience (NIEeS, Cambridge, United Kingdom. This 3rd workshop of the Genomic Standards Consortium was divided into two parts. The first half of the three-day workshop was dedicated to reviewing the genomic diversity of our current and future genome and metagenome collection, and exploring linkages to a series of existing projects through formal presentations. The second half was dedicated to strategic discussions. Outcomes of the workshop include a revised “Minimum Information about a Genome Sequence” (MIGS specification (v1.1, consensus on a variety of features to be added to the Genome Catalogue (GCat, agreement by several researchers to adopt MIGS for imminent genome publications, and an agreement by the EBI and NCBI to input their genome collections into GCat for the purpose of quantifying the amount of optional data already available (e.g., for geographic location coordinates and working towards a single, global list of all public genomes and metagenomes.

  19. Geo-spatial Service and Application based on National E-government Network Platform and Cloud

    Science.gov (United States)

    Meng, X.; Deng, Y.; Li, H.; Yao, L.; Shi, J.

    2014-04-01

    With the acceleration of China's informatization process, our party and government take a substantive stride in advancing development and application of digital technology, which promotes the evolution of e-government and its informatization. Meanwhile, as a service mode based on innovative resources, cloud computing may connect huge pools together to provide a variety of IT services, and has become one relatively mature technical pattern with further studies and massive practical applications. Based on cloud computing technology and national e-government network platform, "National Natural Resources and Geospatial Database (NRGD)" project integrated and transformed natural resources and geospatial information dispersed in various sectors and regions, established logically unified and physically dispersed fundamental database and developed national integrated information database system supporting main e-government applications. Cross-sector e-government applications and services are realized to provide long-term, stable and standardized natural resources and geospatial fundamental information products and services for national egovernment and public users.

  20. Planck 2013 results. XXVIII. The Planck Catalogue of Compact Sources

    DEFF Research Database (Denmark)

    Planck Collaboration,; Ade, P. A. R.; Aghanim, N.

    2013-01-01

    The Planck Catalogue of Compact Sources (PCCS) is the catalogue of sources detected in the Planck nominal mission data. It consists of nine single-frequency catalogues of compact sources containing reliable sources, both Galactic and extragalactic, detected over the entire sky. The PCCS covers th...

  1. CHALLENGES AND OPPORTUNITIES OF CATALOGUE RETAILING

    OpenAIRE

    Heri Bezic; Katija Vojvodic; Zrinka Gjanovic

    2012-01-01

    Today`s retail environment is characterised by new, store and non-store, retailing formats, a wide range of new products, the use of new information and communication technologies and, consequently, the changing customer behaviour. Catalogue retailing is a non-store retail format that has a long history in North America and Europe. Previous research revealed that the primary shopping motives related to catalogue retailing were convenience oriented. Other motives included recreational orientat...

  2. Planck 2013 results. XXIX. Planck catalogue of Sunyaev-Zeldovich sources

    CERN Document Server

    Ade, P.A.R.; Armitage-Caplan, C.; Arnaud, M.; Ashdown, M.; Atrio-Barandela, F.; Aumont, J.; Aussel, H.; Baccigalupi, C.; Banday, A.J.; Barreiro, R.B.; Barrena, R.; Bartelmann, M.; Bartlett, J.G.; Battaner, E.; Benabed, K.; Benoit, A.; Benoit-Levy, A.; Bernard, J.P.; Bersanelli, M.; Bielewicz, P.; Bikmaev, I.; Bobin, J.; Bock, J.J.; Bohringer, H.; Bonaldi, A.; Bond, J.R.; Borrill, J.; Bouchet, F.R.; Bridges, M.; Bucher, M.; Burenin, R.; Burigana, C.; Butler, R.C.; Cardoso, J.F.; Carvalho, P.; Catalano, A.; Challinor, A.; Chamballu, A.; Chary, R.R.; Chen, X.; Chiang, L.Y.; Chiang, H.C.; Chon, G.; Christensen, P.R.; Churazov, E.; Church, S.; Clements, D.L.; Colombi, S.; Colombo, L.P.L.; Comis, B.; Couchot, F.; Coulais, A.; Crill, B.P.; Curto, A.; Cuttaia, F.; Da Silva, A.; Dahle, H.; Danese, L.; Davies, R.D.; Davis, R.J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Delouis, J.M.; Democles, J.; Desert, F.X.; Dickinson, C.; Diego, J.M.; Dolag, K.; Dole, H.; Donzelli, S.; Dore, O.; Douspis, M.; Dupac, X.; Efstathiou, G.; Ensslin, T.A.; Eriksen, H.K.; Feroz, F.; Finelli, F.; Flores-Cacho, I.; Forni, O.; Frailis, M.; Franceschi, E.; Fromenteau, S.; Galeotta, S.; Ganga, K.; Genova-Santos, R.T.; Giard, M.; Giardino, G.; Gilfanov, M.; Giraud-Heraud, Y.; Gonzalez-Nuevo, J.; Gorski, K.M.; Grainge, K.J.B.; Gratton, S.; Gregorio, A.; N, E.Groeneboom; Gruppuso, A.; Hansen, F.K.; Hanson, D.; Harrison, D.; Hempel, A.; Henrot-Versille, S.; Hernandez-Monteagudo, C.; Herranz, D.; Hildebrandt, S.R.; Hivon, E.; Hobson, M.; Holmes, W.A.; Hornstrup, A.; Hovest, W.; Huffenberger, K.M.; Hurier, G.; Hurley-Walker, N.; Jaffe, T.R.; Jaffe, A.H.; Jones, W.C.; Juvela, M.; Keihanen, E.; Keskitalo, R.; Khamitov, I.; Kisner, T.S.; Kneissl, R.; Knoche, J.; Knox, L.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lahteenmaki, A.; Lamarre, J.M.; Lasenby, A.; Laureijs, R.J.; Lawrence, C.R.; Leahy, J.P.; Leonardi, R.; Leon-Tavares, J.; Lesgourgues, J.; Li, C.; Liddle, A.; Liguori, M.; Lilje, P.B.; Linden-Vornle, M.; Lopez-Caniego, M.; Lubin, P.M.; Macias-Perez, J.F.; MacTavish, C.J.; Maffei, B.; Maino, D.; Mandolesi, N.; Maris, M.; Marshall, D.J.; Martin, P.G.; Martinez-Gonzalez, E.; Masi, S.; Matarrese, S.; Matthai, F.; Mazzotta, P.; Mei, S.; Meinhold, P.R.; Melchiorri, A.; Melin, J.B.; Mendes, L.; Mennella, A.; Migliaccio, M.; Mitra, S.; Miville-Deschenes, M.A.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Munshi, D.; Naselsky, P.; Nati, F.; Natoli, P.; Nesvadba, N.P.H.; Netterfield, C.B.; Norgaard-Nielsen, H.U.; Noviello, F.; Novikov, D.; Novikov, I.; O'Dwyer, I.J.; Olamaie, M.; Osborne, S.; Oxborrow, C.A.; Paci, F.; Pagano, L.; Pajot, F.; Paoletti, D.; Pasian, F.; Patanchon, G.; Pearson, T.J.; Perdereau, O.; Perotto, L.; Perrott, Y.C.; Perrotta, F.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Ponthieu, N.; Popa, L.; Poutanen, T.; Pratt, G.W.; Prezeau, G.; Prunet, S.; Puget, J.L.; Rachen, J.P.; Reach, W.T.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Ricciardi, S.; Riller, T.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Roudier, G.; Rowan-Robinson, M.; Rubino-Martin, J.A.; Rumsey, C.; Rusholme, B.; Sandri, M.; Santos, D.; Saunders, R.D.E.; Savini, G.; Scott, D.; Seiffert, M.D.; Shellard, E.P.S.; Shimwell, T.W.; Spencer, L.D.; Starck, J.L.; Stolyarov, V.; Stompor, R.; Sudiwala, R.; Sunyaev, R.; Sureau, F.; Sutton, D.; Suur-Uski, A.S.; Sygnet, J.F.; Tauber, J.A.; Tavagnacco, D.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tristram, M.; Tucci, M.; Tuovinen, J.; Turler, M.; Umana, G.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Vibert, L.; Vielva, P.; Villa, F.; Vittorio, N.; Wade, L.A.; Wandelt, B.D.; White, M.; White, S.D.M.; Yvon, D.; Zacchei, A.; Zonca, A.

    2014-01-01

    We describe the all-sky Planck catalogue of clusters and cluster candidates derived from Sunyaev--Zeldovich (SZ) effect detections using the first 15.5 months of Planck satellite observations. The catalogue contains 1227 entries, making it over six times the size of the Planck Early SZ (ESZ) sample and the largest SZ-selected catalogue to date. It contains 861 confirmed clusters, of which 178 have been confirmed as clusters, mostly through follow-up observations, and a further 683 are previously-known clusters. The remaining 366 have the status of cluster candidates, and we divide them into three classes according to the quality of evidence that they are likely to be true clusters. The Planck SZ catalogue is the deepest all-sky cluster catalogue, with redshifts up to about one, and spans the broadest cluster mass range from (0.1 to 1.6) 10^{15}Msun. Confirmation of cluster candidates through comparison with existing surveys or cluster catalogues is extensively described, as is the statistical characterization...

  3. Dark Energy Survey Year 1 Results: Weak Lensing Shape Catalogues

    Energy Technology Data Exchange (ETDEWEB)

    Zuntz, J.; et al.

    2017-08-04

    We present two galaxy shape catalogues from the Dark Energy Survey Year 1 data set, covering 1500 square degrees with a median redshift of $0.59$. The catalogues cover two main fields: Stripe 82, and an area overlapping the South Pole Telescope survey region. We describe our data analysis process and in particular our shape measurement using two independent shear measurement pipelines, METACALIBRATION and IM3SHAPE. The METACALIBRATION catalogue uses a Gaussian model with an innovative internal calibration scheme, and was applied to $riz$-bands, yielding 34.8M objects. The IM3SHAPE catalogue uses a maximum-likelihood bulge/disc model calibrated using simulations, and was applied to $r$-band data, yielding 21.9M objects. Both catalogues pass a suite of null tests that demonstrate their fitness for use in weak lensing science. We estimate the 1$\\sigma$ uncertainties in multiplicative shear calibration to be $0.013$ and $0.025$ for the METACALIBRATION and IM3SHAPE catalogues, respectively.

  4. GSKY: A scalable distributed geospatial data server on the cloud

    Science.gov (United States)

    Rozas Larraondo, Pablo; Pringle, Sean; Antony, Joseph; Evans, Ben

    2017-04-01

    Earth systems, environmental and geophysical datasets are an extremely valuable sources of information about the state and evolution of the Earth. Being able to combine information coming from different geospatial collections is in increasing demand by the scientific community, and requires managing and manipulating data with different formats and performing operations such as map reprojections, resampling and other transformations. Due to the large data volume inherent in these collections, storing multiple copies of them is unfeasible and so such data manipulation must be performed on-the-fly using efficient, high performance techniques. Ideally this should be performed using a trusted data service and common system libraries to ensure wide use and reproducibility. Recent developments in distributed computing based on dynamic access to significant cloud infrastructure opens the door for such new ways of processing geospatial data on demand. The National Computational Infrastructure (NCI), hosted at the Australian National University (ANU), has over 10 Petabytes of nationally significant research data collections. Some of these collections, which comprise a variety of observed and modelled geospatial data, are now made available via a highly distributed geospatial data server, called GSKY (pronounced [jee-skee]). GSKY supports on demand processing of large geospatial data products such as satellite earth observation data as well as numerical weather products, allowing interactive exploration and analysis of the data. It dynamically and efficiently distributes the required computations among cloud nodes providing a scalable analysis framework that can adapt to serve large number of concurrent users. Typical geospatial workflows handling different file formats and data types, or blending data in different coordinate projections and spatio-temporal resolutions, is handled transparently by GSKY. This is achieved by decoupling the data ingestion and indexing process as

  5. Dynamic Server-Based KML Code Generator Method for Level-of-Detail Traversal of Geospatial Data

    Science.gov (United States)

    Baxes, Gregory; Mixon, Brian; Linger, TIm

    2013-01-01

    Web-based geospatial client applications such as Google Earth and NASA World Wind must listen to data requests, access appropriate stored data, and compile a data response to the requesting client application. This process occurs repeatedly to support multiple client requests and application instances. Newer Web-based geospatial clients also provide user-interactive functionality that is dependent on fast and efficient server responses. With massively large datasets, server-client interaction can become severely impeded because the server must determine the best way to assemble data to meet the client applications request. In client applications such as Google Earth, the user interactively wanders through the data using visually guided panning and zooming actions. With these actions, the client application is continually issuing data requests to the server without knowledge of the server s data structure or extraction/assembly paradigm. A method for efficiently controlling the networked access of a Web-based geospatial browser to server-based datasets in particular, massively sized datasets has been developed. The method specifically uses the Keyhole Markup Language (KML), an Open Geospatial Consortium (OGS) standard used by Google Earth and other KML-compliant geospatial client applications. The innovation is based on establishing a dynamic cascading KML strategy that is initiated by a KML launch file provided by a data server host to a Google Earth or similar KMLcompliant geospatial client application user. Upon execution, the launch KML code issues a request for image data covering an initial geographic region. The server responds with the requested data along with subsequent dynamically generated KML code that directs the client application to make follow-on requests for higher level of detail (LOD) imagery to replace the initial imagery as the user navigates into the dataset. The approach provides an efficient data traversal path and mechanism that can be

  6. Matilda, where are you: subject description of juvenile fiction in the Slovenian catalogue and catalogues of neighbouring countries

    Directory of Open Access Journals (Sweden)

    Alenka Šauperl

    2009-01-01

    Full Text Available Differences in subject description of juvenile fiction was investigated on five examples of international classics in five library catalogues: Oton Župančič Public Library (Knjižnica Otona Župančiča in Ljubljana, Slovenia, Stadtbibliothek public library in Graz, Austria, integrated catalogues of libraries in the Gorizia region in Italy (Sistema bibliotecario della Provincia di Gorizia and the Karlovac region in Croatia (Skupni katalog knjižnica Karlovačke županije in September 2008. As Slovenian youth rarely speaks languages of neighbouring countries, British Library catalogue was added.Results show that catalogue records are inconsistent within an individual library as well as in comparision with other libraries in the sample. Librarians do not make consistent subject descriptions. Class number, which is present in all catalogues except in the Austrian one, usually represents: the author’s country, language and/or nationality,the literary genre, and the target audience.Subject headings in the sample bring information on the subject (aboutness, author’s country, language and/or nationality, the literary genre, and target audience. Summaries tell more on the story. But they can also bring information on emotional experience of the reader, information on the author or history of the literary work. It would be economically beneficial if subject description could be more consistent. But uniform subject description is not possible because of diverse library collections and users.The solution might be in the use of multiple levels of subject description regarding to the type of the libraries.

  7. The Raincoast eCatalogue: the creation of an electronic catalogue as a supplemental selling tool for sales representatives

    OpenAIRE

    Kemp, Elizabeth Anne

    2011-01-01

    Raincoast Books Distribution Ltd. is a Canadian book distributor that provides sales, marketing and distribution services for a number of international and Canadian publishers. Each publishing season Raincoast Books distributes approximately 25,000 paper catalogues to sales representatives and retail accounts. Traditional paper catalogues have major disadvantages including their static format, high cost of production and distribution, inclusion of frontlist titles only and environmental impac...

  8. Integration of Geospatial Science in Teacher Education

    Science.gov (United States)

    Hauselt, Peggy; Helzer, Jennifer

    2012-01-01

    One of the primary missions of our university is to train future primary and secondary teachers. Geospatial sciences, including GIS, have long been excluded from teacher education curriculum. This article explains the curriculum revisions undertaken to increase the geospatial technology education of future teachers. A general education class…

  9. A Spatial Data Infrastructure Integrating Multisource Heterogeneous Geospatial Data and Time Series: A Study Case in Agriculture

    Directory of Open Access Journals (Sweden)

    Gloria Bordogna

    2016-05-01

    Full Text Available Currently, the best practice to support land planning calls for the development of Spatial Data Infrastructures (SDI capable of integrating both geospatial datasets and time series information from multiple sources, e.g., multitemporal satellite data and Volunteered Geographic Information (VGI. This paper describes an original OGC standard interoperable SDI architecture and a geospatial data and metadata workflow for creating and managing multisource heterogeneous geospatial datasets and time series, and discusses it in the framework of the Space4Agri project study case developed to support the agricultural sector in Lombardy region, Northern Italy. The main novel contributions go beyond the application domain for which the SDI has been developed and are the following: the ingestion within an a-centric SDI, potentially distributed in several nodes on the Internet to support scalability, of products derived by processing remote sensing images, authoritative data, georeferenced in-situ measurements and voluntary information (VGI created by farmers and agronomists using an original Smart App; the workflow automation for publishing sets and time series of heterogeneous multisource geospatial data and relative web services; and, finally, the project geoportal, that can ease the analysis of the geospatial datasets and time series by providing complex intelligent spatio-temporal query and answering facilities.

  10. Geospatial Data Analysis Facility

    Data.gov (United States)

    Federal Laboratory Consortium — Geospatial application development, location-based services, spatial modeling, and spatial analysis are examples of the many research applications that this facility...

  11. Bridging the Gap Between Surveyors and the Geo-Spatial Society

    Science.gov (United States)

    Müller, H.

    2016-06-01

    For many years FIG, the International Association of Surveyors, has been trying to bridge the gap between surveyors and the geospatial society as a whole, with the geospatial industries in particular. Traditionally the surveying profession contributed to the good of society by creating and maintaining highly precise and accurate geospatial data bases, based on an in-depth knowledge of spatial reference frameworks. Furthermore in many countries surveyors may be entitled to make decisions about land divisions and boundaries. By managing information spatially surveyors today develop into the role of geo-data managers, the longer the more. Job assignments in this context include data entry management, data and process quality management, design of formal and informal systems, information management, consultancy, land management, all that in close cooperation with many different stakeholders. Future tasks will include the integration of geospatial information into e-government and e-commerce systems. The list of professional tasks underpins the capabilities of surveyors to contribute to a high quality geospatial data and information management. In that way modern surveyors support the needs of a geo-spatial society. The paper discusses several approaches to define the role of the surveyor within the modern geospatial society.

  12. Geospatial field applications within United States Department of Agriculture, Veterinary Services.

    Science.gov (United States)

    FitzMaurice, Priscilla L; Freier, Jerome E; Geter, Kenneth D

    2007-01-01

    Epidemiologists, veterinary medical officers and animal health technicians within Veterinary Services (VS) are actively utilising global positioning system (GPS) technology to obtain positional data on livestock and poultry operations throughout the United States. Geospatial data, if acquired for monitoring and surveillance purposes, are stored within the VS Generic Database (GDB). If the information is collected in response to an animal disease outbreak, the data are entered into the Emergency Management Response System (EMRS). The Spatial Epidemiology group within the Centers for Epidemiology and Animal Health (CEAH) has established minimum data accuracy standards for geodata acquisition. To ensure that field-collected geographic coordinates meet these minimum standards, field personnel are trained in proper data collection procedures. Positional accuracy is validated with digital atlases, aerial photographs, Web-based parcel maps, or address geocoding. Several geospatial methods and technologies are under investigation for future use within VS. These include the direct transfer of coordinates from GPS receivers to computers, GPS-enabled digital cameras, tablet PCs, and GPS receivers preloaded with custom ArcGIS maps - all with the objective of reducing transcription and data entry errors and improving the ease of data collection in the field.

  13. Automatic geospatial information Web service composition based on ontology interface matching

    Science.gov (United States)

    Xu, Xianbin; Wu, Qunyong; Wang, Qinmin

    2008-10-01

    With Web services technology the functions of WebGIS can be presented as a kind of geospatial information service, and helped to overcome the limitation of the information-isolated situation in geospatial information sharing field. Thus Geospatial Information Web service composition, which conglomerates outsourced services working in tandem to offer value-added service, plays the key role in fully taking advantage of geospatial information services. This paper proposes an automatic geospatial information web service composition algorithm that employed the ontology dictionary WordNet to analyze semantic distances among the interfaces. Through making matching between input/output parameters and the semantic meaning of pairs of service interfaces, a geospatial information web service chain can be created from a number of candidate services. A practice of the algorithm is also proposed and the result of it shows the feasibility of this algorithm and the great promise in the emerging demand for geospatial information web service composition.

  14. Geospatial Information from Satellite Imagery for Geovisualisation of Smart Cities in India

    Science.gov (United States)

    Mohan, M.

    2016-06-01

    In the recent past, there have been large emphasis on extraction of geospatial information from satellite imagery. The Geospatial information are being processed through geospatial technologies which are playing important roles in developing of smart cities, particularly in developing countries of the world like India. The study is based on the latest geospatial satellite imagery available for the multi-date, multi-stage, multi-sensor, and multi-resolution. In addition to this, the latest geospatial technologies have been used for digital image processing of remote sensing satellite imagery and the latest geographic information systems as 3-D GeoVisualisation, geospatial digital mapping and geospatial analysis for developing of smart cities in India. The Geospatial information obtained from RS and GPS systems have complex structure involving space, time and presentation. Such information helps in 3-Dimensional digital modelling for smart cities which involves of spatial and non-spatial information integration for geographic visualisation of smart cites in context to the real world. In other words, the geospatial database provides platform for the information visualisation which is also known as geovisualisation. So, as a result there have been an increasing research interest which are being directed to geospatial analysis, digital mapping, geovisualisation, monitoring and developing of smart cities using geospatial technologies. However, the present research has made an attempt for development of cities in real world scenario particulary to help local, regional and state level planners and policy makers to better understand and address issues attributed to cities using the geospatial information from satellite imagery for geovisualisation of Smart Cities in emerging and developing country, India.

  15. Brokered virtual hubs for facilitating access and use of geospatial Open Data

    Science.gov (United States)

    Mazzetti, Paolo; Latre, Miguel; Kamali, Nargess; Brumana, Raffaella; Braumann, Stefan; Nativi, Stefano

    2016-04-01

    Open Data is a major trend in current information technology scenario and it is often publicised as one of the pillars of the information society in the near future. In particular, geospatial Open Data have a huge potential also for Earth Sciences, through the enablement of innovative applications and services integrating heterogeneous information. However, open does not mean usable. As it was recognized at the very beginning of the Web revolution, many different degrees of openness exist: from simple sharing in a proprietary format to advanced sharing in standard formats and including semantic information. Therefore, to fully unleash the potential of geospatial Open Data, advanced infrastructures are needed to increase the data openness degree, enhancing their usability. In October 2014, the ENERGIC OD (European NEtwork for Redistributing Geospatial Information to user Communities - Open Data) project, funded by the European Union under the Competitiveness and Innovation framework Programme (CIP), has started. In response to the EU call, the general objective of the project is to "facilitate the use of open (freely available) geographic data from different sources for the creation of innovative applications and services through the creation of Virtual Hubs". The ENERGIC OD Virtual Hubs aim to facilitate the use of geospatial Open Data by lowering and possibly removing the main barriers which hampers geo-information (GI) usage by end-users and application developers. Data and services heterogeneity is recognized as one of the major barriers to Open Data (re-)use. It imposes end-users and developers to spend a lot of effort in accessing different infrastructures and harmonizing datasets. Such heterogeneity cannot be completely removed through the adoption of standard specifications for service interfaces, metadata and data models, since different infrastructures adopt different standards to answer to specific challenges and to address specific use-cases. Thus

  16. Broad Absorption Line Quasar catalogues with Supervised Neural Networks

    International Nuclear Information System (INIS)

    Scaringi, Simone; Knigge, Christian; Cottis, Christopher E.; Goad, Michael R.

    2008-01-01

    We have applied a Learning Vector Quantization (LVQ) algorithm to SDSS DR5 quasar spectra in order to create a large catalogue of broad absorption line quasars (BALQSOs). We first discuss the problems with BALQSO catalogues constructed using the conventional balnicity and/or absorption indices (BI and AI), and then describe the supervised LVQ network we have trained to recognise BALQSOs. The resulting BALQSO catalogue should be substantially more robust and complete than BI-or AI-based ones.

  17. Mapping a Difference: The Power of Geospatial Visualization

    Science.gov (United States)

    Kolvoord, B.

    2015-12-01

    Geospatial Technologies (GST), such as GIS, GPS and remote sensing, offer students and teachers the opportunity to study the "why" of where. By making maps and collecting location-based data, students can pursue authentic problems using sophisticated tools. The proliferation of web- and cloud-based tools has made these technologies broadly accessible to schools. In addition, strong spatial thinking skills have been shown to be a key factor in supporting students that want to study science, technology, engineering, and mathematics (STEM) disciplines (Wai, Lubinski and Benbow) and pursue STEM careers. Geospatial technologies strongly scaffold the development of these spatial thinking skills. For the last ten years, the Geospatial Semester, a unique dual-enrollment partnership between James Madison University and Virginia high schools, has provided students with the opportunity to use GST's to hone their spatial thinking skills and to do extended projects of local interest, including environmental, geological and ecological studies. Along with strong spatial thinking skills, these students have also shown strong problem solving skills, often beyond those of fellow students in AP classes. Programs like the Geospatial Semester are scalable and within the reach of many college and university departments, allowing strong engagement with K-12 schools. In this presentation, we'll share details of the Geospatial Semester and research results on the impact of the use of these technologies on students' spatial thinking skills, and discuss the success and challenges of developing K-12 partnerships centered on geospatial visualization.

  18. GEOSPATIAL INFORMATION FROM SATELLITE IMAGERY FOR GEOVISUALISATION OF SMART CITIES IN INDIA

    Directory of Open Access Journals (Sweden)

    M. Mohan

    2016-06-01

    Full Text Available In the recent past, there have been large emphasis on extraction of geospatial information from satellite imagery. The Geospatial information are being processed through geospatial technologies which are playing important roles in developing of smart cities, particularly in developing countries of the world like India. The study is based on the latest geospatial satellite imagery available for the multi-date, multi-stage, multi-sensor, and multi-resolution. In addition to this, the latest geospatial technologies have been used for digital image processing of remote sensing satellite imagery and the latest geographic information systems as 3-D GeoVisualisation, geospatial digital mapping and geospatial analysis for developing of smart cities in India. The Geospatial information obtained from RS and GPS systems have complex structure involving space, time and presentation. Such information helps in 3-Dimensional digital modelling for smart cities which involves of spatial and non-spatial information integration for geographic visualisation of smart cites in context to the real world. In other words, the geospatial database provides platform for the information visualisation which is also known as geovisualisation. So, as a result there have been an increasing research interest which are being directed to geospatial analysis, digital mapping, geovisualisation, monitoring and developing of smart cities using geospatial technologies. However, the present research has made an attempt for development of cities in real world scenario particulary to help local, regional and state level planners and policy makers to better understand and address issues attributed to cities using the geospatial information from satellite imagery for geovisualisation of Smart Cities in emerging and developing country, India.

  19. GeoSpatial Data Analysis for DHS Programs

    Energy Technology Data Exchange (ETDEWEB)

    Stephan, Eric G.; Burke, John S.; Carlson, Carrie A.; Gillen, David S.; Joslyn, Cliff A.; Olsen, Bryan K.; Critchlow, Terence J.

    2009-05-10

    The Department of Homeland Security law enforcement faces the continual challenge of analyzing their custom data sources in a geospatial context. From a strategic perspective law enforcement has certain requirements to first broadly characterize a given situation using their custom data sources and then once it is summarily understood, to geospatially analyze their data in detail.

  20. International Atomic Energy Agency Publications. Catalogue 1986-1999

    International Nuclear Information System (INIS)

    2000-11-01

    This catalogue lists all sales publications of the International Atomic Energy Agency issued from 1986 up to the end of 1999 and still available. Some earlier titles which form part of an established series or are still considered important have also been included. The catalogue is in CD-ROM format

  1. Automated geospatial Web Services composition based on geodata quality requirements

    Science.gov (United States)

    Cruz, Sérgio A. B.; Monteiro, Antonio M. V.; Santos, Rafael

    2012-10-01

    Service-Oriented Architecture and Web Services technologies improve the performance of activities involved in geospatial analysis with a distributed computing architecture. However, the design of the geospatial analysis process on this platform, by combining component Web Services, presents some open issues. The automated construction of these compositions represents an important research topic. Some approaches to solving this problem are based on AI planning methods coupled with semantic service descriptions. This work presents a new approach using AI planning methods to improve the robustness of the produced geospatial Web Services composition. For this purpose, we use semantic descriptions of geospatial data quality requirements in a rule-based form. These rules allow the semantic annotation of geospatial data and, coupled with the conditional planning method, this approach represents more precisely the situations of nonconformities with geodata quality that may occur during the execution of the Web Service composition. The service compositions produced by this method are more robust, thus improving process reliability when working with a composition of chained geospatial Web Services.

  2. 9 CFR 3.7 - Compatible grouping.

    Science.gov (United States)

    2010-01-01

    ... 9 Animals and Animal Products 1 2010-01-01 2010-01-01 false Compatible grouping. 3.7 Section 3.7... Cats 1 Animal Health and Husbandry Standards § 3.7 Compatible grouping. Dogs and cats that are housed...; (c) Puppies or kittens 4 months of age or less may not be housed in the same primary enclosure with...

  3. Catalogue of meteorites from South America

    CERN Document Server

    Acevedo, Rogelio Daniel; García, Víctor Manuel

    2014-01-01

    The first Catalogue of Meteorites from South America includes new specimens never previously reported, while doubtful cases and pseudometeorites have been deliberately omitted.The falling of these objects is a random event, but the sites where old meteorites are found tend to be focused in certain areas, e.g. in the deflation surfaces in Chile's Atacama Desert, due to favorable climate conditions and ablation processes.Our Catalogue provides basic information on each specimen like its provenance and the place where it was discovered (in geographic co-ordinates and with illustrative maps), its

  4. OpenSearch technology for geospatial resources discovery

    Science.gov (United States)

    Papeschi, Fabrizio; Enrico, Boldrini; Mazzetti, Paolo

    2010-05-01

    set of services for discovery, access, and processing of geospatial resources in a SOA framework. GI-cat is a distributed CSW framework implementation developed by the ESSI Lab of the Italian National Research Council (CNR-IMAA) and the University of Florence. It provides brokering and mediation functionalities towards heterogeneous resources and inventories, exposing several standard interfaces for query distribution. This work focuses on a new GI-cat interface which allows the catalog to be queried according to the OpenSearch syntax specification, thus filling the gap between the SOA architectural design of the CSW and the Web 2.0. At the moment, there is no OGC standard specification about this topic, but an official change request has been proposed in order to enable the OGC catalogues to support OpenSearch queries. In this change request, an OpenSearch extension is proposed providing a standard mechanism to query a resource based on temporal and geographic extents. Two new catalog operations are also proposed, in order to publish a suitable OpenSearch interface. This extended interface is implemented by the modular GI-cat architecture adding a new profiling module called "OpenSearch profiler". Since GI-cat also acts as a clearinghouse catalog, another component called "OpenSearch accessor" is added in order to access OpenSearch compliant services. An important role in the GI-cat extension, is played by the adopted mapping strategy. Two different kind of mappings are required: query, and response elements mapping. Query mapping is provided in order to fit the simple OpenSearch query syntax to the complex CSW query expressed by the OGC Filter syntax. GI-cat internal data model is based on the ISO-19115 profile, that is more complex than the simple XML syndication formats, such as RSS 2.0 and Atom 1.0, suggested by OpenSearch. Once response elements are available, in order to be presented, they need to be translated from the GI-cat internal data model, to the above

  5. High Resolution Dsm and Classified Volumetric Generation: AN Operational Approach to the Improvement of Geospatial Intelligence

    Science.gov (United States)

    Boccardo, P.; Gentili, G.

    2011-09-01

    As mentioned by Bacastow and Bellafiore, Geospatial Intelligence (GEOINT) is a field of knowledge, a process, and a profession. As knowledge, it is information integrated in a coherent space-time context that supports descriptions, explanations, or forecasts of human activities with which decision makers take action. As a process, it is the means by which data and information are collected, manipulated, geospatially reasoned, and disseminated to decision-makers. The geospatial intelligence professional establishes the scope of activities, interdisciplinary associations, competencies, and standards in academe, government, and the private sectors. Taking into account the fact that GEOINT is crucial for broad organizations, BLOM Group, a leading International provider within acquisition, processing and modeling of geographic information and ITHACA, a non-profit organization devoted to products and services delivering to the UN System in the field of geomatics, set up and provided GEOINT data to the main Italian companies operating in the field of mobile phone networking. This data, extremely useful for telecom network planning, have derived and produced using a standardized and effective (from the production point of view) approach. In this paper, all the procedures used for the production are described and tested with the aim to investigate the suitability of the data and the procedures themselves to any others possible fields of application.

  6. Capacity Building through Geospatial Education in Planning and School Curricula

    Science.gov (United States)

    Kumar, P.; Siddiqui, A.; Gupta, K.; Jain, S.; Krishna Murthy, Y. V. N.

    2014-11-01

    Geospatial technology has widespread usage in development planning and resource management. It offers pragmatic tools to help urban and regional planners to realize their goals. On the request of Ministry of Urban Development, Govt. of India, the Indian Institute of Remote Sensing (IIRS), Dehradun has taken an initiative to study the model syllabi of All India Council for Technical Education for planning curricula of Bachelor and Master (five disciplines) programmes. It is inferred that geospatial content across the semesters in various planning fields needs revision. It is also realized that students pursuing planning curricula are invariably exposed to spatial mapping tools but the popular digital drafting software have limitations on geospatial analysis of planning phenomena. Therefore, students need exposure on geospatial technologies to understand various real world phenomena. Inputs were given to seamlessly merge and incorporate geospatial components throughout the semesters wherever seems relevant. Another initiative by IIRS was taken to enhance the understanding and essence of space and geospatial technologies amongst the young minds at 10+2 level. The content was proposed in a manner such that youngsters start realizing the innumerable contributions made by space and geospatial technologies in their day-to-day life. This effort both at school and college level would help in not only enhancing job opportunities for young generation but also utilizing the untapped human resource potential. In the era of smart cities, higher economic growth and aspirations for a better tomorrow, integration of Geospatial technologies with conventional wisdom can no longer be ignored.

  7. Challenges in sharing of geospatial data by data custodians in South Africa

    Science.gov (United States)

    Kay, Sissiel E.

    2018-05-01

    As most development planning and rendering of public services happens at a place or in a space, geospatial data is required. This geospatial data is best managed through a spatial data infrastructure, which has as a key objective to share geospatial data. The collection and maintenance of geospatial data is expensive and time consuming and so the principle of "collect once - use many times" should apply. It is best to obtain the geospatial data from the authoritative source - the appointed data custodian. In South Africa the South African Spatial Data Infrastructure (SASDI) is the means to achieve the requirement for geospatial data sharing. This requires geospatial data sharing to take place between the data custodian and the user. All data custodians are expected to comply with the Spatial Data Infrastructure Act (SDI Act) in terms of geo-spatial data sharing. Currently data custodians are experiencing challenges with regard to the sharing of geospatial data. This research is based on the current ten data themes selected by the Committee for Spatial Information and the organisations identified as the data custodians for these ten data themes. The objectives are to determine whether the identified data custodians comply with the SDI Act with respect to geospatial data sharing, and if not what are the reasons for this. Through an international comparative assessment it then determines if the compliance with the SDI Act is not too onerous on the data custodians. The research concludes that there are challenges with geospatial data sharing in South Africa and that the data custodians only partially comply with the SDI Act in terms of geospatial data sharing. However, it is shown that the South African legislation is not too onerous on the data custodians.

  8. An Assessment of Online Public Access Catalogue (OPAC ...

    African Journals Online (AJOL)

    The main purpose of this study was to assess the computerized catalogue and its utilization in university libraries in Lagos state. Survey research method was employed for the study. The population for the study was drawn from two university libraries in Lagos state that have automated their catalogues. These libraries are ...

  9. The ASAS-SN bright supernova catalogue - III. 2016

    DEFF Research Database (Denmark)

    Holoien, T. W. -S.; Brown, J. S.; Stanek, K. Z.

    2017-01-01

    This catalogue summarizes information for all supernovae discovered by the All-Sky Automated Survey for SuperNovae (ASAS-SN) and all other bright (m(peak)d......This catalogue summarizes information for all supernovae discovered by the All-Sky Automated Survey for SuperNovae (ASAS-SN) and all other bright (m(peak)d...

  10. Gamma radiography of defects in concrete. Catalogue of reference exposures

    International Nuclear Information System (INIS)

    1974-01-01

    A catalogue of reference exposure as a basic document for the interpretation of gamma-radiographs of reinforced and prestressed concrete structures is presented. The radiation sources are Iridium 192, Caesium 137 and Cobalt 60. Photographic films are used as radiation detectors combined with intensifying screens and filters of lead, copper and iron. The concrete specimens were designed and made for the purpose of studying, as a function of the concrete thickness, the possibility of detecting certain characteristic incorporations or defects. Each set of standard specimens consists of seven standard blocks with the dimensions 15x15x50cm. The thicknesses of the specimens range from 15 to 75cm (1 to 5 blocks)

  11. The geospatial data quality REST API for primary biodiversity data.

    Science.gov (United States)

    Otegui, Javier; Guralnick, Robert P

    2016-06-01

    We present a REST web service to assess the geospatial quality of primary biodiversity data. It enables access to basic and advanced functions to detect completeness and consistency issues as well as general errors in the provided record or set of records. The API uses JSON for data interchange and efficient parallelization techniques for fast assessments of large datasets. The Geospatial Data Quality API is part of the VertNet set of APIs. It can be accessed at http://api-geospatial.vertnet-portal.appspot.com/geospatial and is already implemented in the VertNet data portal for quality reporting. Source code is freely available under GPL license from http://www.github.com/vertnet/api-geospatial javier.otegui@gmail.com or rguralnick@flmnh.ufl.edu Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.

  12. The AKARI IRC asteroid flux catalogue: updated diameters and albedos

    Science.gov (United States)

    Alí-Lagoa, V.; Müller, T. G.; Usui, F.; Hasegawa, S.

    2018-05-01

    The AKARI IRC all-sky survey provided more than twenty thousand thermal infrared observations of over five thousand asteroids. Diameters and albedos were obtained by fitting an empirically calibrated version of the standard thermal model to these data. After the publication of the flux catalogue in October 2016, our aim here is to present the AKARI IRC all-sky survey data and discuss valuable scientific applications in the field of small body physical properties studies. As an example, we update the catalogue of asteroid diameters and albedos based on AKARI using the near-Earth asteroid thermal model (NEATM). We fit the NEATM to derive asteroid diameters and, whenever possible, infrared beaming parameters. We fit groups of observations taken for the same object at different epochs of the survey separately, so we compute more than one diameter for approximately half of the catalogue. We obtained a total of 8097 diameters and albedos for 5170 asteroids, and we fitted the beaming parameter for almost two thousand of them. When it was not possible to fit the beaming parameter, we used a straight line fit to our sample's beaming parameter-versus-phase angle plot to set the default value for each fit individually instead of using a single average value. Our diameters agree with stellar-occultation-based diameters well within the accuracy expected for the model. They also match the previous AKARI-based catalogue at phase angles lower than 50°, but we find a systematic deviation at higher phase angles, at which near-Earth and Mars-crossing asteroids were observed. The AKARI IRC All-sky survey is an essential source of information about asteroids, especially the large ones, since, it provides observations at different observation geometries, rotational coverages and aspect angles. For example, by comparing in more detail a few asteroids for which dimensions were derived from occultations, we discuss how the multiple observations per object may already provide three

  13. VIRAC: the VVV Infrared Astrometric Catalogue

    Science.gov (United States)

    Smith, L. C.; Lucas, P. W.; Kurtev, R.; Smart, R.; Minniti, D.; Borissova, J.; Jones, H. R. A.; Zhang, Z. H.; Marocco, F.; Contreras Peña, C.; Gromadzki, M.; Kuhn, M. A.; Drew, J. E.; Pinfield, D. J.; Bedin, L. R.

    2018-02-01

    We present VIRAC version 1, a near-infrared proper motion and parallax catalogue of the VISTA Variables in the Via Lactea (VVV) survey for 312 587 642 unique sources averaged across all overlapping pawprint and tile images covering 560 deg2 of the bulge of the Milky Way and southern disc. The catalogue includes 119 million high-quality proper motion measurements, of which 47 million have statistical uncertainties below 1 mas yr-1. In the 11 stars and brown dwarfs, subdwarfs, white dwarfs) and kinematic distance measurements of young stellar objects. Nearby objects discovered include LTT 7251 B, an L7 benchmark companion to a G dwarf with over 20 published elemental abundances, a bright L subdwarf, VVV 1256-6202, with extremely blue colours and nine new members of the 25 pc sample. We also demonstrate why this catalogue remains useful in the era of Gaia. Future versions will be based on profile fitting photometry, use the Gaia absolute reference frame and incorporate the longer time baseline of the VVV extended survey.

  14. Establishment of the Northeast Coastal Watershed Geospatial Data Network (NECWGDN)

    Energy Technology Data Exchange (ETDEWEB)

    Hannigan, Robyn [University of Massachusetts Boston

    2014-02-17

    The goals of NECWGDN were to establish integrated geospatial databases that interfaced with existing open-source (water.html) environmental data server technologies (e.g., HydroDesktop) and included ecological and human data to enable evaluation, prediction, and adaptation in coastal environments to climate- and human-induced threats to the coastal marine resources within the Gulf of Maine. We have completed the development and testing of a "test bed" architecture that is compatible with HydroDesktop and have identified key metadata structures that will enable seamless integration and delivery of environmental, ecological, and human data as well as models to predict threats to end-users. Uniquely this database integrates point as well as model data and so offers capacities to end-users that are unique among databases. Future efforts will focus on the development of integrated environmental-human dimension models that can serve, in near real time, visualizations of threats to coastal resources and habitats.

  15. Leveraging geospatial data, technology, and methods for improving the health of communities: priorities and strategies from an expert panel convened by the CDC.

    Science.gov (United States)

    Elmore, Kim; Flanagan, Barry; Jones, Nicholas F; Heitgerd, Janet L

    2010-04-01

    In 2008, CDC convened an expert panel to gather input on the use of geospatial science in surveillance, research and program activities focused on CDC's Healthy Communities Goal. The panel suggested six priorities: spatially enable and strengthen public health surveillance infrastructure; develop metrics for geospatial categorization of community health and health inequity; evaluate the feasibility and validity of standard metrics of community health and health inequities; support and develop GIScience and geospatial analysis; provide geospatial capacity building, training and education; and, engage non-traditional partners. Following the meeting, the strategies and action items suggested by the expert panel were reviewed by a CDC subcommittee to determine priorities relative to ongoing CDC geospatial activities, recognizing that many activities may need to occur either in parallel, or occur multiple times across phases. Phase A of the action items centers on developing leadership support. Phase B focuses on developing internal and external capacity in both physical (e.g., software and hardware) and intellectual infrastructure. Phase C of the action items plan concerns the development and integration of geospatial methods. In summary, the panel members provided critical input to the development of CDC's strategic thinking on integrating geospatial methods and research issues across program efforts in support of its Healthy Communities Goal.

  16. A resource-oriented architecture for a Geospatial Web

    Science.gov (United States)

    Mazzetti, Paolo; Nativi, Stefano

    2010-05-01

    In this presentation we discuss some architectural issues on the design of an architecture for a Geospatial Web, that is an information system for sharing geospatial resources according to the Web paradigm. The success of the Web in building a multi-purpose information space, has raised questions about the possibility of adopting the same approach for systems dedicated to the sharing of more specific resources, such as the geospatial information, that is information characterized by spatial/temporal reference. To this aim an investigation on the nature of the Web and on the validity of its paradigm for geospatial resources is required. The Web was born in the early 90's to provide "a shared information space through which people and machines could communicate" [Berners-Lee 1996]. It was originally built around a small set of specifications (e.g. URI, HTTP, HTML, etc.); however, in the last two decades several other technologies and specifications have been introduced in order to extend its capabilities. Most of them (e.g. the SOAP family) actually aimed to transform the Web in a generic Distributed Computing Infrastructure. While these efforts were definitely successful enabling the adoption of service-oriented approaches for machine-to-machine interactions supporting complex business processes (e.g. for e-Government and e-Business applications), they do not fit in the original concept of the Web. In the year 2000, R. T. Fielding, one of the designers of the original Web specifications, proposes a new architectural style for distributed systems, called REST (Representational State Transfer), aiming to capture the fundamental characteristics of the Web as it was originally conceived [Fielding 2000]. In this view, the nature of the Web lies not so much in the technologies, as in the way they are used. Maintaining the Web architecture conform to the REST style would then assure the scalability, extensibility and low entry barrier of the original Web. On the contrary

  17. Innovative maintenance concept 0f Siemens Energy Service. Turbine documentation in an online catalogue; Innovatives Instandhaltungskonzept bei Siemens Energy Service. Turbinendokumentation als Onlinekatalog

    Energy Technology Data Exchange (ETDEWEB)

    Bitzer, Gerd [Wessendorf Software and Consulting, Berlin (Germany); Stass, Georg [Tema Technologie Marketing AG, Berlin (Germany)

    2010-10-04

    For optimization of its maintenance services, Siemens Energy Services has drawn up an online catalogue as a standardized solution for turbine documentation for customers and service technicians. The new catalogue offers power plant operators internet access to the current documentation of service-relevant replacement parts for machinery. The catalogues are based on SAP data using the Linkone software; each of them has up to 3,000 material items and 1,000 illustrations of replacement parts depending on the type of turbine. The project partner was the Berlin consulting service Wessendorf Software + Consulting (WSC) who have vast experience in the implementation of international catalogue and documentation projects with SAP software. After a project period of 15 months, the innovative solution was implemented in April 2010 and then launched in the market. (orig.)

  18. Integrating Free and Open Source Solutions into Geospatial Science Education

    Directory of Open Access Journals (Sweden)

    Vaclav Petras

    2015-06-01

    Full Text Available While free and open source software becomes increasingly important in geospatial research and industry, open science perspectives are generally less reflected in universities’ educational programs. We present an example of how free and open source software can be incorporated into geospatial education to promote open and reproducible science. Since 2008 graduate students at North Carolina State University have the opportunity to take a course on geospatial modeling and analysis that is taught with both proprietary and free and open source software. In this course, students perform geospatial tasks simultaneously in the proprietary package ArcGIS and the free and open source package GRASS GIS. By ensuring that students learn to distinguish between geospatial concepts and software specifics, students become more flexible and stronger spatial thinkers when choosing solutions for their independent work in the future. We also discuss ways to continually update and improve our publicly available teaching materials for reuse by teachers, self-learners and other members of the GIS community. Only when free and open source software is fully integrated into geospatial education, we will be able to encourage a culture of openness and, thus, enable greater reproducibility in research and development applications.

  19. Geospatial Absorption and Regional Effects

    Directory of Open Access Journals (Sweden)

    IOAN MAC

    2009-01-01

    Full Text Available The geospatial absorptions are characterized by a specific complexity both in content and in their phenomenological and spatial manifestation fields. Such processes are differentiated according to their specificity to pre-absorption, absorption or post-absorption. The mechanisms that contribute to absorption are extremely numerous: aggregation, extension, diffusion, substitution, resistivity (resilience, stratification, borrowings, etc. Between these mechanisms frequent relations are established determining an amplification of the process and of its regional effects. The installation of the geographic osmosis phenomenon in a given territory (a place for example leads to a homogenization of the geospatial state and to the installation of the regional homogeneity.

  20. Planck 2015 results. XXVI. The Second Planck Catalogue of Compact Sources

    CERN Document Server

    Ade, P.A.R.; Argueso, F.; Arnaud, M.; Ashdown, M.; Aumont, J.; Baccigalupi, C.; Banday, A.J.; Barreiro, R.B.; Bartolo, N.; Battaner, E.; Beichman, C.; Benabed, K.; Benoit, A.; Benoit-Levy, A.; Bernard, J.P.; Bersanelli, M.; Bielewicz, P.; Bock, J.J.; Bohringer, H.; Bonaldi, A.; Bonavera, L.; Bond, J.R.; Borrill, J.; Bouchet, F.R.; Boulanger, F.; Bucher, M.; Burigana, C.; Butler, R.C.; Calabrese, E.; Cardoso, J.F.; Carvalho, P.; Catalano, A.; Challinor, A.; Chamballu, A.; Chary, R.R.; Chiang, H.C.; Christensen, P.R.; Clemens, M.; Clements, D.L.; Colombi, S.; Colombo, L.P.L.; Combet, C.; Couchot, F.; Coulais, A.; Crill, B.P.; Curto, A.; Cuttaia, F.; Danese, L.; Davies, R.D.; Davis, R.J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Desert, F.X.; Dickinson, C.; Diego, J.M.; Dole, H.; Donzelli, S.; Dore, O.; Douspis, M.; Ducout, A.; Dupac, X.; Efstathiou, G.; Elsner, F.; Ensslin, T.A.; Eriksen, H.K.; Falgarone, E.; Fergusson, J.; Finelli, F.; Forni, O.; Frailis, M.; Fraisse, A.A.; Franceschi, E.; Frejsel, A.; Galeotta, S.; Galli, S.; Ganga, K.; Giard, M.; Giraud-Heraud, Y.; Gjerlow, E.; Gonzalez-Nuevo, J.; Gorski, K.M.; Gratton, S.; Gregorio, A.; Gruppuso, A.; Gudmundsson, J.E.; Hansen, F.K.; Hanson, D.; Harrison, D.L.; Helou, G.; Henrot-Versille, S.; Hernandez-Monteagudo, C.; Herranz, D.; Hildebrandt, S.R.; Hivon, E.; Hobson, M.; Holmes, W.A.; Hornstrup, A.; Hovest, W.; Huffenberger, K.M.; Hurier, G.; Jaffe, A.H.; Jaffe, T.R.; Jones, W.C.; Juvela, M.; Keihanen, E.; Keskitalo, R.; Kisner, T.S.; Kneissl, R.; Knoche, J.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lahteenmaki, A.; Lamarre, J.M.; Lasenby, A.; Lattanzi, M.; Lawrence, C.R.; Leahy, J.P.; Leonardi, R.; Leon-Tavares, J.; Lesgourgues, J.; Levrier, F.; Liguori, M.; Lilje, P.B.; Linden-Vornle, M.; Lopez-Caniego, M.; Lubin, P.M.; Macias-Perez, J.F.; Maggio, G.; Maino, D.; Mandolesi, N.; Mangilli, A.; Maris, M.; Marshall, D.J.; Martin, P.G.; Martinez-Gonzalez, E.; Masi, S.; Matarrese, S.; McGehee, P.; Meinhold, P.R.; Melchiorri, A.; Mendes, L.; Mennella, A.; Migliaccio, M.; Mitra, S.; Miville-Deschenes, M.A.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Moss, A.; Munshi, D.; Murphy, J.A.; Naselsky, P.; Nati, F.; Natoli, P.; Negrello, M.; Netterfield, C.B.; Norgaard-Nielsen, H.U.; Noviello, F.; Novikov, D.; Novikov, I.; Oxborrow, C.A.; Paci, F.; Pagano, L.; Pajot, F.; Paladini, R.; Paoletti, D.; Partridge, B.; Pasian, F.; Patanchon, G.; Pearson, T.J.; Perdereau, O.; Perotto, L.; Perrotta, F.; Pettorino, V.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Pratt, G.W.; Prezeau, G.; Prunet, S.; Puget, J.L.; Rachen, J.P.; Reach, W.T.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Renzi, A.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Rossetti, M.; Roudier, G.; Rowan-Robinson, M.; Rubino-Martin, J.A.; Rusholme, B.; Sandri, M.; Sanghera, H.S.; Santos, D.; Savelainen, M.; Savini, G.; Scott, D.; Seiffert, M.D.; Shellard, E.P.S.; Spencer, L.D.; Stolyarov, V.; Sudiwala, R.; Sunyaev, R.; Sutton, D.; Suur-Uski, A.S.; Sygnet, J.F.; Tauber, J.A.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tornikoski, M.; Tristram, M.; Tucci, M.; Tuovinen, J.; Turler, M.; Umana, G.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Vielva, P.; Villa, F.; Wade, L.A.; Walter, B.; Wandelt, B.D.; Wehus, I.K.; Yvon, D.; Zacchei, A.; Zonca, A.

    2016-01-01

    The Second Planck Catalogue of Compact Sources is a catalogue of sources detected in single-frequency maps from the full duration of the Planck mission and supersedes previous versions of the Planck compact source catalogues. It consists of compact sources, both Galactic and extragalactic, detected over the entire sky. Compact sources detected in the lower frequency channels are assigned to the PCCS2, while at higher frequencies they are assigned to one of two sub-catalogues, the PCCS2 or PCCS2E, depending on their location on the sky. The first of these catalogues covers most of the sky and allows the user to produce subsamples at higher reliabilities than the target 80% integral reliability of the catalogue. The PCCS2E contains sources detected in sky regions where the diffuse emission makes it difficult to quantify the reliability of the detections. Both the PCCS2 and PCCS2E include polarization measurements, in the form of polarized flux densities, or upper limits, and orientation angles for all seven pol...

  1. Remapping simulated halo catalogues in redshift space

    OpenAIRE

    Mead, Alexander; Peacock, John

    2014-01-01

    We discuss the extension to redshift space of a rescaling algorithm, designed to alter the effective cosmology of a pre-existing simulated particle distribution or catalogue of dark matter haloes. The rescaling approach was initially developed by Angulo & White and was adapted and applied to halo catalogues in real space in our previous work. This algorithm requires no information other than the initial and target cosmological parameters, and it contains no tuned parameters. It is shown here ...

  2. Creation of Defects Catalogue for Nonconforming Product Identification in the Foundry Organization

    Directory of Open Access Journals (Sweden)

    Andrea Sütőová

    2013-12-01

    Full Text Available The paper deals with system of casting defects classification problematics and creation of defects catalogue in the foundry organization. There is described the value of correct defects classification and identification in the literature review and also some tools for defects classification are mentioned. Existing defects classifications and catalogues are often unusable for particular production processes and casting technology. Many foundries therefore create their own defects catalogues. The sample of created catalogue, which classifies and describes defects occuring in the aluminium foundry organization and its benefits are presented in the paper. The created catalogue primarily serves as a visual support for production operators and quality control processes.

  3. Collective Sensing: Integrating Geospatial Technologies to Understand Urban Systems—An Overview

    Directory of Open Access Journals (Sweden)

    Geoffrey J. Hay

    2011-08-01

    Full Text Available Cities are complex systems composed of numerous interacting components that evolve over multiple spatio-temporal scales. Consequently, no single data source is sufficient to satisfy the information needs required to map, monitor, model, and ultimately understand and manage our interaction within such urban systems. Remote sensing technology provides a key data source for mapping such environments, but is not sufficient for fully understanding them. In this article we provide a condensed urban perspective of critical geospatial technologies and techniques: (i Remote Sensing; (ii Geographic Information Systems; (iii object-based image analysis; and (iv sensor webs, and recommend a holistic integration of these technologies within the language of open geospatial consortium (OGC standards in-order to more fully understand urban systems. We then discuss the potential of this integration and conclude that this extends the monitoring and mapping options beyond “hard infrastructure” by addressing “humans as sensors”, mobility and human-environment interactions, and future improvements to quality of life and of social infrastructures.

  4. Biosecurity and geospatial analysis of mycoplasma infections in ...

    African Journals Online (AJOL)

    Geospatial database of farm locations and biosecurity measures are essential to control disease outbreaks. A study was conducted to establish geospatial database on poultry farms in Al-Jabal Al-Gharbi region of Libya, to evaluate the biosecurity level of each farm and to determine the seroprevalence of mycoplasma and ...

  5. Searches over graphs representing geospatial-temporal remote sensing data

    Science.gov (United States)

    Brost, Randolph; Perkins, David Nikolaus

    2018-03-06

    Various technologies pertaining to identifying objects of interest in remote sensing images by searching over geospatial-temporal graph representations are described herein. Graphs are constructed by representing objects in remote sensing images as nodes, and connecting nodes with undirected edges representing either distance or adjacency relationships between objects and directed edges representing changes in time. Geospatial-temporal graph searches are made computationally efficient by taking advantage of characteristics of geospatial-temporal data in remote sensing images through the application of various graph search techniques.

  6. Uniform Title in Theory and in Slovenian and Croatian Cataloguing Practice

    Directory of Open Access Journals (Sweden)

    Marija Petek

    2013-09-01

    Full Text Available ABSTRACTPurpose:  The paper investigates the importance and development of uniform title that enables collocation in the library catalogue. Research results on use of uniform titles in two union catalogues, the Slovenian COBISS and the Croatian CROLIST are also presented.Methodology/approach:  Theoretical apects of the uniform title are treated: for the first time by Panizzi, then in the Paris Principles being the basis for the Verona's cataloguing code; in the latest International Cataloguing Principles including conceptual models Functional Requirements for Bibliographic Records (FRBR and Functional Requirements for Authority Data (FRAD; and last but not least in the international cataloguing code Resource Description and Access (RDA. To find out whether the uniform titles are used consistently according to the Verona's cataloguing code and to the requirements of the bibliographic formats COMARC and UNIMARC, the frequency of tags 300 and 500 in bibliographic records is explored.Results:  The research results indicate that the use of uniform titles in COBISS and CROLIST is not satisfactory and that the tags 300 and 500 are often missing in bibliographic recods. In online catalogues a special attention should be given to the uniform title as it is considered an efficient linking device in the catalogue and as it enables collocation.Research limitations:  The research is limited to bibliographic records for translations of works of personal authors and of anonymous works; corporate authors are not included.Originality/practical implications:  Presenting development of the uniform title from the very beginning up to now and the first research on the uniform title in COBISS.

  7. GISpark: A Geospatial Distributed Computing Platform for Spatiotemporal Big Data

    Science.gov (United States)

    Wang, S.; Zhong, E.; Wang, E.; Zhong, Y.; Cai, W.; Li, S.; Gao, S.

    2016-12-01

    Geospatial data are growing exponentially because of the proliferation of cost effective and ubiquitous positioning technologies such as global remote-sensing satellites and location-based devices. Analyzing large amounts of geospatial data can provide great value for both industrial and scientific applications. Data- and compute- intensive characteristics inherent in geospatial big data increasingly pose great challenges to technologies of data storing, computing and analyzing. Such challenges require a scalable and efficient architecture that can store, query, analyze, and visualize large-scale spatiotemporal data. Therefore, we developed GISpark - a geospatial distributed computing platform for processing large-scale vector, raster and stream data. GISpark is constructed based on the latest virtualized computing infrastructures and distributed computing architecture. OpenStack and Docker are used to build multi-user hosting cloud computing infrastructure for GISpark. The virtual storage systems such as HDFS, Ceph, MongoDB are combined and adopted for spatiotemporal data storage management. Spark-based algorithm framework is developed for efficient parallel computing. Within this framework, SuperMap GIScript and various open-source GIS libraries can be integrated into GISpark. GISpark can also integrated with scientific computing environment (e.g., Anaconda), interactive computing web applications (e.g., Jupyter notebook), and machine learning tools (e.g., TensorFlow/Orange). The associated geospatial facilities of GISpark in conjunction with the scientific computing environment, exploratory spatial data analysis tools, temporal data management and analysis systems make up a powerful geospatial computing tool. GISpark not only provides spatiotemporal big data processing capacity in the geospatial field, but also provides spatiotemporal computational model and advanced geospatial visualization tools that deals with other domains related with spatial property. We

  8. The National 3-D Geospatial Information Web-Based Service of Korea

    Science.gov (United States)

    Lee, D. T.; Kim, C. W.; Kang, I. G.

    2013-09-01

    3D geospatial information systems should provide efficient spatial analysis tools and able to use all capabilities of the third dimension, and a visualization. Currently, many human activities make steps toward the third dimension like land use, urban and landscape planning, cadastre, environmental monitoring, transportation monitoring, real estate market, military applications, etc. To reflect this trend, the Korean government has been started to construct the 3D geospatial data and service platform. Since the geospatial information was introduced in Korea, the construction of geospatial information (3D geospatial information, digital maps, aerial photographs, ortho photographs, etc.) has been led by the central government. The purpose of this study is to introduce the Korean government-lead 3D geospatial information web-based service for the people who interested in this industry and we would like to introduce not only the present conditions of constructed 3D geospatial data but methodologies and applications of 3D geospatial information. About 15% (about 3,278.74 km2) of the total urban area's 3D geospatial data have been constructed by the national geographic information institute (NGII) of Korea from 2005 to 2012. Especially in six metropolitan cities and Dokdo (island belongs to Korea) on level of detail (LOD) 4 which is photo-realistic textured 3D models including corresponding ortho photographs were constructed in 2012. In this paper, we represented web-based 3D map service system composition and infrastructure and comparison of V-world with Google Earth service will be presented. We also represented Open API based service cases and discussed about the protection of location privacy when we construct 3D indoor building models. In order to prevent an invasion of privacy, we processed image blurring, elimination and camouflage. The importance of public-private cooperation and advanced geospatial information policy is emphasized in Korea. Thus, the progress of

  9. Geo-Spatial Support for Assessment of Anthropic Impact on Biodiversity

    Directory of Open Access Journals (Sweden)

    Marco Piragnolo

    2014-04-01

    Full Text Available This paper discusses a methodology where geo-spatial analysis tools are used to quantify risk derived from anthropic activities on habitats and species. The method has been developed with a focus on simplification and the quality of standard procedures set on flora and fauna protected by the European Directives. In this study case, the DPSIR (Drivers, Pressures, State, Impacts, Responses is applied using spatial procedures in a geographical information system (GIS framework. This approach can be inserted in a multidimensional space as the analysis is applied to each threat, pressure and activity and also to each habitat and species, at the spatial and temporal scale. Threats, pressures and activities, stress and indicators can be managed by means of a geo-database and analyzed using spatial analysis functions in a tested GIS workflow environment. The method applies a matrix with risk values, and the final product is a geo-spatial representation of impact indicators, which can be used as a support for decision-makers at various levels (regional, national and European.

  10. Energy research projects in the Nordic countries - catalogue 1983

    International Nuclear Information System (INIS)

    1983-01-01

    The Nordic energy ministers at their meeting February 9, 1982 agreed upon a working plan for the Nordic energy cooperation. As part of this plan a contact group was established in order to maintain coordination and cooperation within the area of energy research and development. This group decided April 1982 to establish a catalogue of energy research projects in the Nordic countries. A pilot catalogue was published in June 1982. The 1983 catalogue gives an up-to-date survey of energy research and development projects in the Nordic countries. About 2125 projects are described, and information is given on investigator(s), performing organization, financing body, funds, and period. The catalogue is prepared by the Nordic energy libraries through their cooperation in Nordic Atomic Libraries Joint Secretariat. The information is also included in the data base Nordic Energy Index (NEI), which is online accessible at I/S Datacentralen, Copenhagen, via EURONET, SCANNET, TYMNET, AND TELENET. (BP)

  11. Towards a Next-Generation Catalogue Cross-Match Service

    Science.gov (United States)

    Pineau, F.; Boch, T.; Derriere, S.; Arches Consortium

    2015-09-01

    We have been developing in the past several catalogue cross-match tools. On one hand the CDS XMatch service (Pineau et al. 2011), able to perform basic but very efficient cross-matches, scalable to the largest catalogues on a single regular server. On the other hand, as part of the European project ARCHES1, we have been developing a generic and flexible tool which performs potentially complex multi-catalogue cross-matches and which computes probabilities of association based on a novel statistical framework. Although the two approaches have been managed so far as different tracks, the need for next generation cross-match services dealing with both efficiency and complexity is becoming pressing with forthcoming projects which will produce huge high quality catalogues. We are addressing this challenge which is both theoretical and technical. In ARCHES we generalize to N catalogues the candidate selection criteria - based on the chi-square distribution - described in Pineau et al. (2011). We formulate and test a number of Bayesian hypothesis which necessarily increases dramatically with the number of catalogues. To assign a probability to each hypotheses, we rely on estimated priors which account for local densities of sources. We validated our developments by comparing the theoretical curves we derived with the results of Monte-Carlo simulations. The current prototype is able to take into account heterogeneous positional errors, object extension and proper motion. The technical complexity is managed by OO programming design patterns and SQL-like functionalities. Large tasks are split into smaller independent pieces for scalability. Performances are achieved resorting to multi-threading, sequential reads and several tree data-structures. In addition to kd-trees, we account for heterogeneous positional errors and object's extension using M-trees. Proper-motions are supported using a modified M-tree we developed, inspired from Time Parametrized R-trees (TPR

  12. Planck 2015 results: XXVIII. The Planck Catalogue of Galactic cold clumps

    DEFF Research Database (Denmark)

    Ade, P. A R; Aghanim, N.; Arnaud, M.

    2016-01-01

    We present the Planck Catalogue of Galactic Cold Clumps (PGCC), an all-sky catalogue of Galactic cold clump candidates detected by Planck. This catalogue is the full version of the Early Cold Core (ECC) catalogue, which was made available in 2011 with the Early Release Compact Source Catalogue (E...

  13. Revision of the tsunami catalogue affecting Turkish coasts and surrounding regions

    Directory of Open Access Journals (Sweden)

    Y. Altinok

    2011-02-01

    Full Text Available The coasts of Turkey have been hit by tsunamis in the past. The first national earthquake-tsunami catalogues were compiled in the early 1980s while the most up-to-date tsunami catalogues are mainly the products of recent European projects. The EU projects GITEC and GITEC-TWO (Genesis and Impact of Tsunamis on the European Coasts and TRANSFER (Tsunami Risk ANd Strategies For the European Region have added important contributions in establishing and developing unified criteria for tsunami parameterisation, standards for the quality of the data, the data format and the database general architecture. On the basis of these new aspects and based on recent marine geophysical data, tsunamigenic earthquakes, tsunami intensities and their reliability have been revised. The current version of the database contains 134 events, most of which have affected the Turkish coasts seriously during the last 3500 years. The reliability index of 76 events was "probable" and "definite", so that they could be used for assessment of the risk along the Turkish coastal region and for implementation of prevention policies.

  14. Revelation of `Hidden' Balinese Geospatial Heritage on A Map

    Science.gov (United States)

    Soeria Atmadja, Dicky A. S.; Wikantika, Ketut; Budi Harto, Agung; Putra, Daffa Gifary M.

    2018-05-01

    Bali is not just about beautiful nature. It also has a unique and interesting cultural heritage, including `hidden' geospatial heritage. Tri Hita Karana is a Hinduism concept of life consisting of human relation to God, to other humans and to the nature (Parahiyangan, Pawongan and Palemahan), Based on it, - in term of geospatial aspect - the Balinese derived its spatial orientation, spatial planning & lay out, measurement as well as color and typography. Introducing these particular heritage would be a very interesting contribution to Bali tourism. As a respond to these issues, a question arise on how to reveal these unique and highly valuable geospatial heritage on a map which can be used to introduce and disseminate them to the tourists. Symbols (patterns & colors), orientation, distance, scale, layout and toponimy have been well known as elements of a map. There is an chance to apply Balinese geospatial heritage in representing these map elements.

  15. Australian comments on data catalogues

    Energy Technology Data Exchange (ETDEWEB)

    Symonds, J L [A.A.E.C. Research Establishment, Lucas Heights (Australia)

    1968-05-01

    Between the need for some neutron data and a final evaluated set of data, the need for an action file, a bibliographic and reference file of catalogue, and a data storage and retrieval file is discussed.

  16. Standard classification: Physics

    International Nuclear Information System (INIS)

    1977-01-01

    This is a draft standard classification of physics. The conception is based on the physics part of the systematic catalogue of the Bayerische Staatsbibliothek and on the classification given in standard textbooks. The ICSU-AB classification now used worldwide by physics information services was not taken into account. (BJ) [de

  17. Bim and Gis: when Parametric Modeling Meets Geospatial Data

    Science.gov (United States)

    Barazzetti, L.; Banfi, F.

    2017-12-01

    Geospatial data have a crucial role in several projects related to infrastructures and land management. GIS software are able to perform advanced geospatial analyses, but they lack several instruments and tools for parametric modelling typically available in BIM. At the same time, BIM software designed for buildings have limited tools to handle geospatial data. As things stand at the moment, BIM and GIS could appear as complementary solutions, notwithstanding research work is currently under development to ensure a better level of interoperability, especially at the scale of the building. On the other hand, the transition from the local (building) scale to the infrastructure (where geospatial data cannot be neglected) has already demonstrated that parametric modelling integrated with geoinformation is a powerful tool to simplify and speed up some phases of the design workflow. This paper reviews such mixed approaches with both simulated and real examples, demonstrating that integration is already a reality at specific scales, which are not dominated by "pure" GIS or BIM. The paper will also demonstrate that some traditional operations carried out with GIS software are also available in parametric modelling software for BIM, such as transformation between reference systems, DEM generation, feature extraction, and geospatial queries. A real case study is illustrated and discussed to show the advantage of a combined use of both technologies. BIM and GIS integration can generate greater usage of geospatial data in the AECOO (Architecture, Engineering, Construction, Owner and Operator) industry, as well as new solutions for parametric modelling with additional geoinformation.

  18. BIM AND GIS: WHEN PARAMETRIC MODELING MEETS GEOSPATIAL DATA

    Directory of Open Access Journals (Sweden)

    L. Barazzetti

    2017-12-01

    Full Text Available Geospatial data have a crucial role in several projects related to infrastructures and land management. GIS software are able to perform advanced geospatial analyses, but they lack several instruments and tools for parametric modelling typically available in BIM. At the same time, BIM software designed for buildings have limited tools to handle geospatial data. As things stand at the moment, BIM and GIS could appear as complementary solutions, notwithstanding research work is currently under development to ensure a better level of interoperability, especially at the scale of the building. On the other hand, the transition from the local (building scale to the infrastructure (where geospatial data cannot be neglected has already demonstrated that parametric modelling integrated with geoinformation is a powerful tool to simplify and speed up some phases of the design workflow. This paper reviews such mixed approaches with both simulated and real examples, demonstrating that integration is already a reality at specific scales, which are not dominated by “pure” GIS or BIM. The paper will also demonstrate that some traditional operations carried out with GIS software are also available in parametric modelling software for BIM, such as transformation between reference systems, DEM generation, feature extraction, and geospatial queries. A real case study is illustrated and discussed to show the advantage of a combined use of both technologies. BIM and GIS integration can generate greater usage of geospatial data in the AECOO (Architecture, Engineering, Construction, Owner and Operator industry, as well as new solutions for parametric modelling with additional geoinformation.

  19. Assessing the catalogue module of Alice for window software ...

    African Journals Online (AJOL)

    The paper presents a general description of Alice For Window Software with a detailed analysis of the catalogue module. It highlights the basic features of the module such as add, edit, delete, search field and the grab button. The cataloguing process is clearly delineated. The paper also discusses Alice For Window ...

  20. A Geospatial Semantic Enrichment and Query Service for Geotagged Photographs

    Science.gov (United States)

    Ennis, Andrew; Nugent, Chris; Morrow, Philip; Chen, Liming; Ioannidis, George; Stan, Alexandru; Rachev, Preslav

    2015-01-01

    With the increasing abundance of technologies and smart devices, equipped with a multitude of sensors for sensing the environment around them, information creation and consumption has now become effortless. This, in particular, is the case for photographs with vast amounts being created and shared every day. For example, at the time of this writing, Instagram users upload 70 million photographs a day. Nevertheless, it still remains a challenge to discover the “right” information for the appropriate purpose. This paper describes an approach to create semantic geospatial metadata for photographs, which can facilitate photograph search and discovery. To achieve this we have developed and implemented a semantic geospatial data model by which a photograph can be enrich with geospatial metadata extracted from several geospatial data sources based on the raw low-level geo-metadata from a smartphone photograph. We present the details of our method and implementation for searching and querying the semantic geospatial metadata repository to enable a user or third party system to find the information they are looking for. PMID:26205265

  1. Analysis of the seismic catalogues for the Vrancea Region, Romania

    International Nuclear Information System (INIS)

    Romashkova, L.L.; Kossobokov, V.G.

    2005-11-01

    Vrancea (Romania) is a geographical region between Eastern and Southern Carpathian Mountains. The region is characterized by a rather high level of seismic activity mainly at intermediate (up to 200 km) depths. These intermediate-depth earthquakes occur between 45 deg-46 deg N and 26 deg-27 deg E. The shallow earthquakes are dispersed over a much broader territory. We performed the comparative analysis of earthquake catalogues available for Vrancea region aiming at the compilation of a data set, to be as complete and homogeneous as possible, which, hopefully, will be used for the prediction of strong and possibly moderate earthquakes in the region by means of M8 algorithm. The two catalogues under study are: 1) Global Hypocenter Data Base catalogue, NEIC (GHDB, 1989) and 2) local Vrancea seismic catalogue (Moldoveanu et al., 1995) and their updates. (author)

  2. An Equivalent Moment Magnitude Earthquake Catalogue for Western Turkey and its Quantitative Properties

    Science.gov (United States)

    Leptokaropoulos, Konstantinos; Vasilios, Karakostas; Eleftheria, Papadimitriou; Aggeliki, Adamaki; Onur, Tan; Zumer, Pabuçcu

    2013-04-01

    Earthquake catalogues consist a basic product of seismology, resulting from complex procedures and suffering from natural and man-made errors. The accumulation of these problems over space and time lead to inhomogeneous catalogues which in turn lead to significant uncertainties in many kinds of analyses, such as seismicity rate evaluation and seismic hazard assessment. A major source of catalogue inhomogeneity is the variety of magnitude scales (i.e. Mw, mb, MS, ML, Md), reported from different institutions and sources. Therefore an effort is made in this study to compile a catalogue as homogenous as possible regarding the magnitude scale for the region of Western Turkey (26oE - 32oE longitude, 35oN - 43oN latitude), one of the most rapidly deforming regions worldwide with intense seismic activity, complex fault systems and frequent strong earthquakes. For this purpose we established new relationships to transform as many as possible available magnitudes into equivalent moment magnitude scale, M*w. These relations yielded by the application of the General Orthogonal Regression method and the statistical significance of the results was quantified. The final equivalent moment magnitude was evaluated by taking into consideration all the available magnitudes for which a relation was obtained and also a weight inversely proportional to their standard deviation. Once the catalogue was compiled the magnitude of completeness, Mc, was investigated in both space and time regime. The b-values and their accuracy were also calculated by the maximum likelihood estimate. The spatial and temporal constraints were selected in respect to seismicity recording level, since the state and evolution of the local and regional seismic networks are unknown. We modified and applied the Goodness of Fit test of Wiemer and Wyss (2000) in order to be more effective in datasets that are characterized by smaller sample size and higher Mcthresholds. The compiled catalogue and the Mcevaluation

  3. A Catalogue of marine biodiversity indicators

    Directory of Open Access Journals (Sweden)

    Heliana Teixeira

    2016-11-01

    Full Text Available A Catalogue of Marine Biodiversity Indicators was developed with the aim of providing the basis for assessing the environmental status of the marine ecosystems. Useful for the implementation of the Marine Strategy Framework Directive (MSFD, this catalogue allows the navigation of a database of indicators mostly related to biological diversity, non-indigenous species, food webs, and seafloor integrity. Over 600 indicators were compiled, which were developed and used in the framework of different initiatives (e.g. EU policies, research projects and in national and international contexts (e.g. Regional Seas Conventions, and assessments in non-European seas. The catalogue reflects the current scientific capability to address environmental assessment needs by providing a broad coverage of the most relevant indicators for marine biodiversity and ecosystem integrity.The available indicators are reviewed according to their typology, data requirements, development status, geographical coverage, relevance to habitats or biodiversity components, and related human pressures. Through this comprehensive overview, we discuss the potential of the current set of indicators in a wide range of contexts, from large-scale to local environmental programs, and we also address shortcomings in light of current needs.Developed by the DEVOTES Project, the catalogue is freely available through the DEVOTool software application, which provides browsing and query options for the associated metadata. The tool allows extraction of ranked indicator lists best fulfilling selected criteria, enabling users to search for suitable indicators to address a particular biodiversity component, ecosystem feature, habitat or pressure in a marine area of interest.This tool is useful for EU Member States, Regional Sea Conventions, the European Commission, non-governmental organizations, managers, scientists and any person interested in marine environmental assessment. It allows users to

  4. A Catalogue of Marine Biodiversity Indicators

    KAUST Repository

    Teixeira, Heliana; Berg, Torsten; Uusitalo, Laura; Fü rhaupter, Karin; Heiskanen, Anna Stiina; Mazik, Krysia; Lynam, Christopher P.; Neville, Suzanna; Rodriguez, J. German; Papadopoulou, Nadia; Moncheva, Snejana; Churilova, Tanya; Kryvenko, Olga; Krause-Jensen, Dorte; Zaiko, Anastasija; Verí ssimo, Helena; Pantazi, Maria; Carvalho, Susana; Patrí cio, Joana; Uyarra, Maria C.; Borja, À ngel

    2016-01-01

    A Catalogue of Marine Biodiversity Indicators was developed with the aim of providing the basis for assessing the environmental status of the marine ecosystems. Useful for the implementation of the Marine Strategy Framework Directive (MSFD), this catalogue allows the navigation of a database of indicators mostly related to biological diversity, non-indigenous species, food webs, and seafloor integrity. Over 600 indicators were compiled, which were developed and used in the framework of different initiatives (e.g., EU policies, research projects) and in national and international contexts (e.g., Regional Seas Conventions, and assessments in non-European seas). The catalogue reflects the current scientific capability to address environmental assessment needs by providing a broad coverage of the most relevant indicators for marine biodiversity and ecosystem integrity. The available indicators are reviewed according to their typology, data requirements, development status, geographical coverage, relevance to habitats or biodiversity components, and related human pressures. Through this comprehensive overview, we discuss the potential of the current set of indicators in a wide range of contexts, from large-scale to local environmental programs, and we also address shortcomings in light of current needs. Developed by the DEVOTES Project, the catalogue is freely available through the DEVOTool software application, which provides browsing and query options for the associated metadata. The tool allows extraction of ranked indicator lists best fulfilling selected criteria, enabling users to search for suitable indicators to address a particular biodiversity component, ecosystem feature, habitat, or pressure in a marine area of interest. This tool is useful for EU Member States, Regional Sea Conventions, the European Commission, non-governmental organizations, managers, scientists, and any person interested in marine environmental assessment. It allows users to build

  5. A Catalogue of Marine Biodiversity Indicators

    KAUST Repository

    Teixeira, Heliana

    2016-11-04

    A Catalogue of Marine Biodiversity Indicators was developed with the aim of providing the basis for assessing the environmental status of the marine ecosystems. Useful for the implementation of the Marine Strategy Framework Directive (MSFD), this catalogue allows the navigation of a database of indicators mostly related to biological diversity, non-indigenous species, food webs, and seafloor integrity. Over 600 indicators were compiled, which were developed and used in the framework of different initiatives (e.g., EU policies, research projects) and in national and international contexts (e.g., Regional Seas Conventions, and assessments in non-European seas). The catalogue reflects the current scientific capability to address environmental assessment needs by providing a broad coverage of the most relevant indicators for marine biodiversity and ecosystem integrity. The available indicators are reviewed according to their typology, data requirements, development status, geographical coverage, relevance to habitats or biodiversity components, and related human pressures. Through this comprehensive overview, we discuss the potential of the current set of indicators in a wide range of contexts, from large-scale to local environmental programs, and we also address shortcomings in light of current needs. Developed by the DEVOTES Project, the catalogue is freely available through the DEVOTool software application, which provides browsing and query options for the associated metadata. The tool allows extraction of ranked indicator lists best fulfilling selected criteria, enabling users to search for suitable indicators to address a particular biodiversity component, ecosystem feature, habitat, or pressure in a marine area of interest. This tool is useful for EU Member States, Regional Sea Conventions, the European Commission, non-governmental organizations, managers, scientists, and any person interested in marine environmental assessment. It allows users to build

  6. Geo3DML: A standard-based exchange format for 3D geological models

    Science.gov (United States)

    Wang, Zhangang; Qu, Honggang; Wu, Zixing; Wang, Xianghong

    2018-01-01

    A geological model (geomodel) in three-dimensional (3D) space is a digital representation of the Earth's subsurface, recognized by geologists and stored in resultant geological data (geodata). The increasing demand for data management and interoperable applications of geomodelscan be addressed by developing standard-based exchange formats for the representation of not only a single geological object, but also holistic geomodels. However, current standards such as GeoSciML cannot incorporate all the geomodel-related information. This paper presents Geo3DML for the exchange of 3D geomodels based on the existing Open Geospatial Consortium (OGC) standards. Geo3DML is based on a unified and formal representation of structural models, attribute models and hierarchical structures of interpreted resultant geodata in different dimensional views, including drills, cross-sections/geomaps and 3D models, which is compatible with the conceptual model of GeoSciML. Geo3DML aims to encode all geomodel-related information integrally in one framework, including the semantic and geometric information of geoobjects and their relationships, as well as visual information. At present, Geo3DML and some supporting tools have been released as a data-exchange standard by the China Geological Survey (CGS).

  7. Derivation of photometric redshifts for the 3XMM catalogue

    Science.gov (United States)

    Georgantopoulos, I.; Corral, A.; Mountrichas, G.; Ruiz, A.; Masoura, V.; Fotopoulou, S.; Watson, M.

    2017-10-01

    We present the results from our ESA Prodex project that aims to derive photometric redshifts for the 3XMM catalogue. The 3XMM DR-6 offers the largest X-ray survey, containing 470,000 unique sources over 1000 sq. degrees. We cross-correlate the X-ray positions with optical and near-IR catalogues using Bayesian statistics. The optical catalogue used so far is the SDSS while currently we are employing the recently released PANSTARRS catalogue. In the near IR we use the Viking, VHS, UKIDS surveys and also the WISE W1 and W2 filters. The estimation of photometric redshifts is based on the TPZ software. The training sample is based on X-ray selected samples with available SDSS spectroscopy. We present here the results for the 40,000 3XMM sources with available SDSS counterparts. Our analysis provides very reliable photometric redshifts with sigma(mad)=0.05 and a fraction of outliers of 8% for the optically extended sources. We discuss the wide range of applications that are feasible using this unprecedented resource.

  8. SWOT analysis on National Common Geospatial Information Service Platform of China

    Science.gov (United States)

    Zheng, Xinyan; He, Biao

    2010-11-01

    Currently, the trend of International Surveying and Mapping is shifting from map production to integrated service of geospatial information, such as GOS of U.S. etc. Under this circumstance, the Surveying and Mapping of China is inevitably shifting from 4D product service to NCGISPC (National Common Geospatial Information Service Platform of China)-centered service. Although State Bureau of Surveying and Mapping of China has already provided a great quantity of geospatial information service to various lines of business, such as emergency and disaster management, transportation, water resource, agriculture etc. The shortcomings of the traditional service mode are more and more obvious, due to the highly emerging requirement of e-government construction, the remarkable development of IT technology and emerging online geospatial service demands of various lines of business. NCGISPC, which aimed to provide multiple authoritative online one-stop geospatial information service and API for further development to government, business and public, is now the strategic core of SBSM (State Bureau of Surveying and Mapping of China). This paper focuses on the paradigm shift that NCGISPC brings up by using SWOT (Strength, Weakness, Opportunity and Threat) analysis, compared to the service mode that based on 4D product. Though NCGISPC is still at its early stage, it represents the future service mode of geospatial information of China, and surely will have great impact not only on the construction of digital China, but also on the way that everyone uses geospatial information service.

  9. DIGI-vis: Distributed interactive geospatial information visualization

    KAUST Repository

    Ponto, Kevin

    2010-03-01

    Geospatial information systems provide an abundance of information for researchers and scientists. Unfortunately this type of data can usually only be analyzed a few megapixels at a time, giving researchers a very narrow view into these voluminous data sets. We propose a distributed data gathering and visualization system that allows researchers to view these data at hundreds of megapixels simultaneously. This system allows scientists to view real-time geospatial information at unprecedented levels expediting analysis, interrogation, and discovery. ©2010 IEEE.

  10. Modeling photovoltaic diffusion: an analysis of geospatial datasets

    International Nuclear Information System (INIS)

    Davidson, Carolyn; Drury, Easan; Lopez, Anthony; Elmore, Ryan; Margolis, Robert

    2014-01-01

    This study combines address-level residential photovoltaic (PV) adoption trends in California with several types of geospatial information—population demographics, housing characteristics, foreclosure rates, solar irradiance, vehicle ownership preferences, and others—to identify which subsets of geospatial information are the best predictors of historical PV adoption. Number of rooms, heating source and house age were key variables that had not been previously explored in the literature, but are consistent with the expected profile of a PV adopter. The strong relationship provided by foreclosure indicators and mortgage status have less of an intuitive connection to PV adoption, but may be highly correlated with characteristics inherent in PV adopters. Next, we explore how these predictive factors and model performance varies between different Investor Owned Utility (IOU) regions in California, and at different spatial scales. Results suggest that models trained with small subsets of geospatial information (five to eight variables) may provide similar explanatory power as models using hundreds of geospatial variables. Further, the predictive performance of models generally decreases at higher resolution, i.e., below ZIP code level since several geospatial variables with coarse native resolution become less useful for representing high resolution variations in PV adoption trends. However, for California we find that model performance improves if parameters are trained at the regional IOU level rather than the state-wide level. We also find that models trained within one IOU region are generally representative for other IOU regions in CA, suggesting that a model trained with data from one state may be applicable in another state. (letter)

  11. Geospatial Services in Special Libraries: A Needs Assessment Perspective

    Science.gov (United States)

    Barnes, Ilana

    2013-01-01

    Once limited to geographers and mapmakers, Geographic Information Systems (GIS) has taken a growing central role in information management and visualization. Geospatial services run a gamut of different products and services from Google maps to ArcGIS servers to Mobile development. Geospatial services are not new. Libraries have been writing about…

  12. An Ontology-supported Approach for Automatic Chaining of Web Services in Geospatial Knowledge Discovery

    Science.gov (United States)

    di, L.; Yue, P.; Yang, W.; Yu, G.

    2006-12-01

    Recent developments in geospatial semantic Web have shown promise for automatic discovery, access, and use of geospatial Web services to quickly and efficiently solve particular application problems. With the semantic Web technology, it is highly feasible to construct intelligent geospatial knowledge systems that can provide answers to many geospatial application questions. A key challenge in constructing such intelligent knowledge system is to automate the creation of a chain or process workflow that involves multiple services and highly diversified data and can generate the answer to a specific question of users. This presentation discusses an approach for automating composition of geospatial Web service chains by employing geospatial semantics described by geospatial ontologies. It shows how ontology-based geospatial semantics are used for enabling the automatic discovery, mediation, and chaining of geospatial Web services. OWL-S is used to represent the geospatial semantics of individual Web services and the type of the services it belongs to and the type of the data it can handle. The hierarchy and classification of service types are described in the service ontology. The hierarchy and classification of data types are presented in the data ontology. For answering users' geospatial questions, an Artificial Intelligent (AI) planning algorithm is used to construct the service chain by using the service and data logics expressed in the ontologies. The chain can be expressed as a graph with nodes representing services and connection weights representing degrees of semantic matching between nodes. The graph is a visual representation of logical geo-processing path for answering users' questions. The graph can be instantiated to a physical service workflow for execution to generate the answer to a user's question. A prototype system, which includes real world geospatial applications, is implemented to demonstrate the concept and approach.

  13. Sustainable energy catalogue - for European decision-makers. Final report

    Energy Technology Data Exchange (ETDEWEB)

    Gram, S.; Jacobsen, Soeren

    2006-10-15

    The Green paper - A European Strategy for Sustainable, Competitive and Secure Energy, 2006 states that Europe has a rising dependency on imported energy reserves, which are concentrated in a few countries. The Rising gas and oil prices along with demands on lower emissions of CO2 adds pressure on the need for a new energy future for Europe. EU has since 1990 planned to become world leader in the renewable energy field. Therefore the EU member states have agreed that by 2010 21% of the consumed electricity and 5,75% of the consumed gasoline and diesel should originate from renewable energy sources. If the EU countries are to reach their goals, a commitment on several levels to develop and install energy from sustainable energy sources is needed. The purpose of this catalogue is to offer planners and decision-makers in EU states an inspirational tool to be used during local or regional transition towards sustainable energy technologies. The catalogue can also be used by everyone else who needs an overview of the sustainable energy technologies and their current development level and future potential, among others educational use is relevant. The catalogue provides an introduction to the following technologies that are already or are estimated to become central to a development with renewable energy in EU: Technologies for wind energy, wave energy, geothermal energy, bioenergy, solar energy, hydropower and fuel cells. The catalogue also includes a section about energy systems, which also includes a part about technologies for efficient use of energy. The catalogue could have included a few other technologies as e.g. heating pumps, but due to the size of the catalogue a priority was necessary. The catalogue does not claim to give all answers or to be complete regarding all details about the individual technologies; even so it offers information, which cannot easily be looked up on the Internet. In the back of the catalogue, under 'References and links' there

  14. Catalogue of theses

    International Nuclear Information System (INIS)

    Paranjpe, S.V.

    1975-01-01

    The catalogue lists 442 theses submitted by the scientists of the Bhabha Atomic Research Centre, since its inception, to the various universities in India and abroad for the award of M. Sc. and Ph. D. degrees. Theses are grouped under broad subject headings which are arranged in the order of Universal Decimal Classification Scheme. In addition to the author and guide index, a detailed subject index is appended which enhances the utility of the compilation. (S.V.P.)

  15. Open Source Web Based Geospatial Processing with OMAR

    Directory of Open Access Journals (Sweden)

    Mark Lucas

    2009-01-01

    Full Text Available The availability of geospatial data sets is exploding. New satellites, aerial platforms, video feeds, global positioning system tagged digital photos, and traditional GIS information are dramatically increasing across the globe. These raw materials need to be dynamically processed, combined and correlated to generate value added information products to answer a wide range of questions. This article provides an overview of OMAR web based geospatial processing. OMAR is part of the Open Source Software Image Map project under the Open Source Geospatial Foundation. The primary contributors of OSSIM make their livings by providing professional services to US Government agencies and programs. OMAR provides one example that open source software solutions are increasingly being deployed in US government agencies. We will also summarize the capabilities of OMAR and its plans for near term development.

  16. The Value of Information - Accounting for a New Geospatial Paradigm

    Science.gov (United States)

    Pearlman, J.; Coote, A. M.

    2014-12-01

    A new frontier in consideration of socio-economic benefit is valuing information as an asset, often referred to as Infonomics. Conventional financial practice does not easily provide a mechanism for valuing information and yet clearly for many of the largest corporations, such as Google and Facebook, it is their principal asset. This is exacerbated for public sector organizations, as those that information-centric rather than information-enabled are relatively few - statistics, archiving and mapping agencies are perhaps the only examples - so it's not at the top of the agenda for Government. However, it is a hugely important issue when valuing Geospatial data and information. Geospatial data allows public institutions to operate, and facilitates the provision of essential services for emergency response and national defense. In this respect, geospatial data is strongly analogous to other types of public infrastructure, such as utilities and roads. The use of Geospatial data is widespread from companies in the transportation or construction sectors to individual planning for daily events. The categorization of geospatial data as infrastructure is critical to decisions related to investment in its management, maintenance and upgrade over time. Geospatial data depreciates in the same way that physical infrastructure depreciates. It needs to be maintained otherwise its functionality and value in use declines. We have coined the term geo-infonomics to encapsulate the concept. This presentation will develop the arguments around its importance and current avenues of research.

  17. A CLOUD-BASED PLATFORM SUPPORTING GEOSPATIAL COLLABORATION FOR GIS EDUCATION

    Directory of Open Access Journals (Sweden)

    X. Cheng

    2015-05-01

    Full Text Available GIS-related education needs support of geo-data and geospatial software. Although there are large amount of geographic information resources distributed on the web, the discovery, process and integration of these resources are still unsolved. Researchers and teachers always searched geo-data by common search engines but results were not satisfied. They also spent much money and energy on purchase and maintenance of various kinds of geospatial software. Aimed at these problems, a cloud-based geospatial collaboration platform called GeoSquare was designed and implemented. The platform serves as a geoportal encouraging geospatial data, information, and knowledge sharing through highly interactive and expressive graphic interfaces. Researchers and teachers can solve their problems effectively in this one-stop solution. Functions, specific design and implementation details are presented in this paper. Site of GeoSquare is: http://geosquare.tianditu.com/

  18. a Cloud-Based Platform Supporting Geospatial Collaboration for GIS Education

    Science.gov (United States)

    Cheng, X.; Gui, Z.; Hu, K.; Gao, S.; Shen, P.; Wu, H.

    2015-05-01

    GIS-related education needs support of geo-data and geospatial software. Although there are large amount of geographic information resources distributed on the web, the discovery, process and integration of these resources are still unsolved. Researchers and teachers always searched geo-data by common search engines but results were not satisfied. They also spent much money and energy on purchase and maintenance of various kinds of geospatial software. Aimed at these problems, a cloud-based geospatial collaboration platform called GeoSquare was designed and implemented. The platform serves as a geoportal encouraging geospatial data, information, and knowledge sharing through highly interactive and expressive graphic interfaces. Researchers and teachers can solve their problems effectively in this one-stop solution. Functions, specific design and implementation details are presented in this paper. Site of GeoSquare is: http://geosquare.tianditu.com/

  19. Netherlands Oil and Gas Catalogue 2010

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2010-04-15

    The Netherlands Oil and Gas Catalogue 2010 is a specialist publication from IRO (the Association of Dutch Suppliers in the Oil and Gas Industry) and Pedemex. A wealth of relevant business information has been collected with the greatest care to be included in this publication, and subsequently brought together in an easy to digest order. The catalogue is broken down into the following headings: (1) Engineering, Consultancy and Research; (2) Exploration, Drilling and Production; (3) Construction and Fabrication; (4) Contracting,Transport and Installation; (5) Equipment Supply; (6) Miscellaneous. In addition you can by using keywords, rapidly identify the company you are looking for. A list is also enclosed with the names and contact details of IRO members, and the sectors in which they are active.

  20. M3.2.3 Personas Catalogue

    DEFF Research Database (Denmark)

    Guldbæk Rasmussen, Katja; Iversen, Rie; Petersen, Gitte

    This catalogue contains 7 personas developed for use in the Europeana projects. The premise of this work has been to find already existing personas within the domains of archives, museums and libraries in Europe. These have then been pared down to their essentials and rebuilt, using input from...... Europeana partners and research on behavior and search patterns. If you have never worked with personas before,please take the time to read the short introduction in the chapter about method. The personas, and a brief “How To” is the central issue in this catalogue and therefore placed at the front....... For those wanting to dig a little deeper into how the personas were created, more in-depth material can be found in the chapters at the back....

  1. Catalogue of ptyctimous mites (Acari, Oribatida) of the world.

    Science.gov (United States)

    NiedbaŁa, Wojciech; Liu, Dong

    2018-03-11

    As important representatives of Oribatida (Acari), ptyctimous mites comprise more than 1400 described species in 40 genera and subgenera, with nearly cosmopolitan distribution except for the Arctic and Antarctic Regions. They are capable of folding the aspidosoma under the opisthosoma to protect their appendages, and are primarily soil and litter inhabitants, feeding on fungi and decaying plant remains with various levels of specificity. Our purpose was to provide a detailed catalogue of all known ptyctimous mite species in the world with information of distribution, taxonomic issues and some remarks. Data of known juvenile  instars of ptyctimous mites which were not included in Norton Ermilov (2014) were added. We hope that our catalogue with bibliography will be helpful in taxonomic and ecological studies.        The catalogue presents taxonomic information and geographic distribution of 1431 known species of the world belonging to 42 genera and eight families (not including data of genus and species inquirenda, nomina nuda and species without author name). Among them, 261 species are listed as synonyms, 43 species inquirenda, nine homonyms, 17 new synonyms, one new subgenus Mahuntritia subgenus nov. and three new names are included in the catalogue.

  2. Version 2000 of the Catalogue of Galactic Planetary Nebulae

    Science.gov (United States)

    Kohoutek, L.

    2001-11-01

    The ``Catalogue of Galactic Planetary Nebulae (Version 2000)'' appears in Abhandlungen aus der Hamburger Sternwarte, Band XII in the year 2001. It is a continuation of CGPN(1967) and contains 1510 objects classified as galactic PNe up to the end of 1999. The lists of possible pre-PNe and possible post-PNe are also given. The catalogue is restricted only to the data belonging to the location and identification of the objects. It gives identification charts of PNe discovered since 1965 (published in the supplements to CGPN) and those charts of objects discovered earlier, which have wrong or uncertain identification. The question ``what is a planetary nebula'' is discussed and the typical values of PNe and of their central stars are summarized. Short statistics about the discoveries of PNe are given. The catalogue is also available in the Centre de Données, Strasbourg and at Hamburg Observatory via internet. The Catalogue is only available in electronic form at the CDS via anonymous ftp to cdsarc.u-strasbg.fr (130.79.128.5) or via http://cdsweb.u-strasbg.fr/cgi-bin/qcat?J/A+A/378/843

  3. INTERACT Station Catalogue - 2015

    DEFF Research Database (Denmark)

    INTERACT stations are located in all major environmental envelopes of the Arctic providing an ideal platform for studying climate change and its impact on the environment and local communities. Since alpine environments face similar changes and challenges as the Arctic, the INTERACT network also ...... catalogue includes descriptions of 73 research stations included in the network at the time of printing....

  4. Planck 2013 results. XXIX. Planck catalogue of Sunyaev-Zeldovich sources

    DEFF Research Database (Denmark)

    Ade, P. A. R.; Aghanim, N.; Armitage-Caplan, C.

    2013-01-01

    We describe the all-sky Planck catalogue of clusters and cluster candidates derived from Sunyaev-Zeldovich (SZ) effect detections using the first 15.5 months of Planck satellite observations. The catalogue contains 1227 entries, making it over six times the size of the Planck Early SZ (ESZ) sampl...

  5. Geo-spatial technologies in urban environments policy, practice, and pixels

    CERN Document Server

    Jensen, Ryan R; McLean, Daniel

    2004-01-01

    Using Geospatial Technologies in Urban Environments simultaneously fills two gaping vacuums in the scholarly literature on urban geography. The first is the clear and straightforward application of geospatial technologies to practical urban issues. By using remote sensing and statistical techniques (correlation-regression analysis, the expansion method, factor analysis, and analysis of variance), the - thors of these 12 chapters contribute significantly to our understanding of how geospatial methodologies enhance urban studies. For example, the GIS Specialty Group of the Association of American Geographers (AAG) has the largest m- bership of all the AAG specialty groups, followed by the Urban Geography S- cialty Group. Moreover, the Urban Geography Specialty Group has the largest number of cross-memberships with the GIS Specialty Group. This book advances this important geospatial and urban link. Second, the book fills a wide void in the urban-environment literature. Although the Annals of the Association of ...

  6. Metadata Wizard: an easy-to-use tool for creating FGDC-CSDGM metadata for geospatial datasets in ESRI ArcGIS Desktop

    Science.gov (United States)

    Ignizio, Drew A.; O'Donnell, Michael S.; Talbert, Colin B.

    2014-01-01

    Creating compliant metadata for scientific data products is mandated for all federal Geographic Information Systems professionals and is a best practice for members of the geospatial data community. However, the complexity of the The Federal Geographic Data Committee’s Content Standards for Digital Geospatial Metadata, the limited availability of easy-to-use tools, and recent changes in the ESRI software environment continue to make metadata creation a challenge. Staff at the U.S. Geological Survey Fort Collins Science Center have developed a Python toolbox for ESRI ArcDesktop to facilitate a semi-automated workflow to create and update metadata records in ESRI’s 10.x software. The U.S. Geological Survey Metadata Wizard tool automatically populates several metadata elements: the spatial reference, spatial extent, geospatial presentation format, vector feature count or raster column/row count, native system/processing environment, and the metadata creation date. Once the software auto-populates these elements, users can easily add attribute definitions and other relevant information in a simple Graphical User Interface. The tool, which offers a simple design free of esoteric metadata language, has the potential to save many government and non-government organizations a significant amount of time and costs by facilitating the development of The Federal Geographic Data Committee’s Content Standards for Digital Geospatial Metadata compliant metadata for ESRI software users. A working version of the tool is now available for ESRI ArcDesktop, version 10.0, 10.1, and 10.2 (downloadable at http:/www.sciencebase.gov/metadatawizard).

  7. Towards Geo-spatial Hypermedia: Concepts and Prototype Implementation

    DEFF Research Database (Denmark)

    Grønbæk, Kaj; Vestergaard, Peter Posselt; Ørbæk, Peter

    2002-01-01

    This paper combines spatial hypermedia with techniques from Geographical Information Systems and location based services. We describe the Topos 3D Spatial Hypermedia system and how it has been developed to support geo-spatial hypermedia coupling hypermedia information to model representations...... of real world buildings and landscapes. The prototype experiments are primarily aimed at supporting architects and landscape architects in their work on site. Here it is useful to be able to superimpose and add different layers of information to, e.g. a landscape depending on the task being worked on. We...... and indirect navigation. Finally, we conclude with a number of research issues which are central to the future development of geo-spatial hypermedia, including design issues in combining metaphorical and literal hypermedia space, as well as a discussion of the role of spatial parsing in a geo-spatial context....

  8. A new reference global instrumental earthquake catalogue (1900-2009)

    Science.gov (United States)

    Di Giacomo, D.; Engdahl, B.; Bondar, I.; Storchak, D. A.; Villasenor, A.; Bormann, P.; Lee, W.; Dando, B.; Harris, J.

    2011-12-01

    For seismic hazard studies on a global and/or regional scale, accurate knowledge of the spatial distribution of seismicity, the magnitude-frequency relation and the maximum magnitudes is of fundamental importance. However, such information is normally not homogeneous (or not available) for the various seismically active regions of the Earth. To achieve the GEM objectives (www.globalquakemodel.org) of calculating and communicating earthquake risk worldwide, an improved reference global instrumental catalogue for large earthquakes spanning the entire 100+ years period of instrumental seismology is an absolute necessity. To accomplish this task, we apply the most up-to-date techniques and standard observatory practices for computing the earthquake location and magnitude. In particular, the re-location procedure benefits both from the depth determination according to Engdahl and Villaseñor (2002), and the advanced technique recently implemented at the ISC (Bondár and Storchak, 2011) to account for correlated error structure. With regard to magnitude, starting from the re-located hypocenters, the classical surface and body-wave magnitudes are determined following the new IASPEI standards and by using amplitude-period data of phases collected from historical station bulletins (up to 1970), which were not available in digital format before the beginning of this work. Finally, the catalogue will provide moment magnitude values (including uncertainty) for each seismic event via seismic moment, via surface wave magnitude or via other magnitude types using empirical relationships. References Engdahl, E.R., and A. Villaseñor (2002). Global seismicity: 1900-1999. In: International Handbook of Earthquake and Engineering Seismology, eds. W.H.K. Lee, H. Kanamori, J.C. Jennings, and C. Kisslinger, Part A, 665-690, Academic Press, San Diego. Bondár, I., and D. Storchak (2011). Improved location procedures at the International Seismological Centre, Geophys. J. Int., doi:10.1111/j

  9. Using the Geospatial Web to Deliver and Teach Giscience Education Programs

    Science.gov (United States)

    Veenendaal, B.

    2015-05-01

    Geographic information science (GIScience) education has undergone enormous changes over the past years. One major factor influencing this change is the role of the geospatial web in GIScience. In addition to the use of the web for enabling and enhancing GIScience education, it is also used as the infrastructure for communicating and collaborating among geospatial data and users. The web becomes both the means and the content for a geospatial education program. However, the web does not replace the traditional face-to-face environment, but rather is a means to enhance it, expand it and enable an authentic and real world learning environment. This paper outlines the use of the web in both the delivery and content of the GIScience program at Curtin University. The teaching of the geospatial web, web and cloud based mapping, and geospatial web services are key components of the program, and the use of the web and online learning are important to deliver this program. Some examples of authentic and real world learning environments are provided including joint learning activities with partner universities.

  10. Where do we go with Union Catalogues?

    Directory of Open Access Journals (Sweden)

    Edmund Chamberlain

    2013-07-01

    Full Text Available The United Kingdom boasts union catalogues for its major research libraries, journal holdings, archives and, most recently, for its public library collections. For researchers wanting to locate material across the UK, such aggregations have long served as a first stop for researchers wanting to find the right material and also provided a showcase for our formidable research collections. In the global networked environment, search engines and social networks can fulfil much of the functionality of union catalogues and have become the natural places to which our users go for search and discovery, even in academic situations. Right now, there is a ‘disconnect’ between the data describing our collections and the places users first turn to start their searches. This can be fixed by exposing descriptive data to wider audiences beyond the silo of the local catalogue, but data publishing is a fast moving area with little obvious short-term institutional-level gain and some start-up barriers. Publishing library data to the open web at the level of a national aggregation would utilize existing skill sets and infrastructure, minimize risk and maximize impact.

  11. HR diagrams derived from the Michigan Spectral Catalogue

    International Nuclear Information System (INIS)

    Houk, N.; Fesen, R.

    1978-01-01

    The authors present some HR diagrams constructed using data from the Michigan Spectral Catalogues. Houk (1975) has been systematically reclassifying the Henry Draper stars on the MK system, from the south pole northward. Objective-prism plates, with a reciprocal dispersion of 108 A/mm, have been taken with the Michigan Curtis Schmidt telescope at Cerro Tololo Inter-American Observatory in Chile. The spectra are classified visually from the plates, and the results are put onto IBM cards and magnetic tape from which the catalogues are produced. (Auth.)

  12. VizieR Online Data Catalog: A unified supernova catalogue (Lennarz+, 2012)

    Science.gov (United States)

    Lennarz, D.; Altmann, D.; Wiebusch, C.

    2011-11-01

    A supernova catalogue containing data for 5526 extragalactic supernovae that were discovered up to 2010 December 31. It combines several catalogues that are currently available online in a consistent and traceable way. During the comparison of the catalogues inconsistent entries were identified and resolved where possible. Remaining inconsistencies are marked transparently and can be easily identified. Thus it is possible to select a high-quality sample in a most simple way. Where available, redshift-based distance estimates to the supernovae were replaced by journal-refereed distances. (1 data file).

  13. Points of View: Herbert Bayer’s Exhibition Catalogue for the 1930 Section Allemande

    Directory of Open Access Journals (Sweden)

    Wallis Miller

    2017-01-01

    Full Text Available Sigfried Giedion called Herbert Bayer’s exhibition catalogue for the 1930 'Section Allemande' a “minor typographical masterpiece.” Like similar catalogues, it is inexpensive, provides an inventory list, has an introduction, functions as a guide, and is illustrated. However, the majority of its images are of installations, not their contents. Bayer accommodates the catalogue type for applied arts exhibitions by listing installations as objects, but he confronts the type by showing installations as display contexts that establish points of view, emulating, idealizing and interpreting the experience of the exhibition. By independently constructing ways of seeing and understanding the exhibition, the catalogue resists being an appendage to the exhibition, despite their close relationship. Giedion may have viewed Bayer’s catalogue as an important but secondary work of graphic design, but this article argues that it is of primary significance as an exhibition catalogue, an unusual essay on the book typology that is conscious of its history while moving outside — to other types of book design and to exhibitions — to transform it.

  14. Selling pictures: the illustrated auction catalogue

    Directory of Open Access Journals (Sweden)

    Elizabeth Pergam

    2014-12-01

    Full Text Available This essay is based upon a survey of reproductions in auction catalogues – from their first appearance in the early eighteenth century until their more consistent use in the second decade of the twentieth century. Examining the role of these illustrations sheds light on how auctions functioned; it was not just the works of art that were traded, but knowledge about those works of art became currency to be exchanged. In contrast to the high end engravings and photographs of luxury illustrated art books, reproductions in auction catalogues – publications produced as ephemeral marketing tools – were of noticeably lower quality. This study of the status of reproductions, therefore, investigates the evolving understanding of art knowledge, both aesthetic and economic, and the interdependence of the market and connoisseurship.

  15. Leveraging the geospatial advantage

    Science.gov (United States)

    Ben Butler; Andrew Bailey

    2013-01-01

    The Wildland Fire Decision Support System (WFDSS) web-based application leverages geospatial data to inform strategic decisions on wildland fires. A specialized data team, working within the Wildland Fire Management Research Development and Application group (WFM RD&A), assembles authoritative national-level data sets defining values to be protected. The use of...

  16. The development of a new edition of the gamma-ray spectrum catalogues designed for presentation in electronic format

    International Nuclear Information System (INIS)

    Heath, R.L.

    1997-01-01

    New editions of the original Gamma-ray Spectrum Catalogues are being prepared for publication in electronic format. The objective of this program is to produce versions of the Catalogues in CD-ROM format and as an Internet resource. Additions to the original content of the Catalogues will include integrated decay scheme drawings, tables of related decay data, and updated text on the techniques of gamma-ray spectrometry. Related decay data from the Evaluated Nuclear Structure Data File (ENSDF) are then added, and all data converted to the Adobe Acrobat (PDF) format for CD-ROM production and availability on the Internet. At a later date the catalogues will be expanded to include spectra representing the response of large-volume Ge detectors, alpha-particle spectra, prompt neutron capture and inelastic scattering gamma-ray spectra, and gross fission product spectra characteristic of fuel cycle waste materials. Characterization of radioactivity in materials is a requirement in many phases of radioactive waste management. Movement, shipping, treatment, all activities which involve handling of mixed waste or TRU categories of waste at all DOE sites will require that measurements and assessment documentation utilize basic nuclear data which are tracable to internationally accepted standard values. This program will involve the identification of data needs unique to the development and application of specialized detector systems for radioactive waste characterization

  17. National Geospatial-Intelligence Agency Academic Research Program

    Science.gov (United States)

    Loomer, S. A.

    2004-12-01

    "Know the Earth.Show the Way." In fulfillment of its vision, the National Geospatial-Intelligence Agency (NGA) provides geospatial intelligence in all its forms and from whatever source-imagery, imagery intelligence, and geospatial data and information-to ensure the knowledge foundation for planning, decision, and action. To achieve this, NGA conducts a multi-disciplinary program of basic research in geospatial intelligence topics through grants and fellowships to the leading investigators, research universities, and colleges of the nation. This research provides the fundamental science support to NGA's applied and advanced research programs. The major components of the NGA Academic Research Program (NARP) are: - NGA University Research Initiatives (NURI): Three-year basic research grants awarded competitively to the best investigators across the US academic community. Topics are selected to provide the scientific basis for advanced and applied research in NGA core disciplines. - Historically Black College and University - Minority Institution Research Initiatives (HBCU-MI): Two-year basic research grants awarded competitively to the best investigators at Historically Black Colleges and Universities, and Minority Institutions across the US academic community. - Director of Central Intelligence Post-Doctoral Research Fellowships: Fellowships providing access to advanced research in science and technology applicable to the intelligence community's mission. The program provides a pool of researchers to support future intelligence community needs and develops long-term relationships with researchers as they move into career positions. This paper provides information about the NGA Academic Research Program, the projects it supports and how other researchers and institutions can apply for grants under the program.

  18. Catalogue of European earthquakes with intensities higher than 4

    International Nuclear Information System (INIS)

    Van Gils, J.M.; Leydecker, G.

    1991-01-01

    The catalogue of European earthquakes with intensities higher than 4 contains some 20 000 seismic events that happened in member countries of the European Communities, Switzerland and Austria. It was prepared on the basis of already existing national catalogues and includes historical data as well as present-day data. All historical data are harmonized as far as possible to the same intensity scale (MSK-scale) to make them suitable for computerization. Present-day data include instrumental and macroseismic data. Instrumental data are expressed in terms of magnitude (Richter scale) while macroseismic data are given in intensities. Compilation of seismic data can provide a basis for statistically supported studies of site selection procedures and the qualitative assessment of seismic risks. Three groups of seismic maps illustrate the content of the catalogue for different time periods and different intensities

  19. Geospatial data infrastructure: The development of metadata for geo-information in China

    Science.gov (United States)

    Xu, Baiquan; Yan, Shiqiang; Wang, Qianju; Lian, Jian; Wu, Xiaoping; Ding, Keyong

    2014-03-01

    Stores of geoscience records are in constant flux. These stores are continually added to by new information, ideas and data, which are frequently revised. The geoscience record is in restrained by human thought and technology for handling information. Conventional methods strive, with limited success, to maintain geoscience records which are readily susceptible and renewable. The information system must adapt to the diversity of ideas and data in geoscience and their changes through time. In China, more than 400,000 types of important geological data are collected and produced in geological work during the last two decades, including oil, natural gas and marine data, mine exploration, geophysical, geochemical, remote sensing and important local geological survey and research reports. Numerous geospatial databases are formed and stored in National Geological Archives (NGA) with available formats of MapGIS, ArcGIS, ArcINFO, Metalfile, Raster, SQL Server, Access and JPEG. But there is no effective way to warrant that the quality of information is adequate in theory and practice for decision making. The need for fast, reliable, accurate and up-to-date information by providing the Geographic Information System (GIS) communities are becoming insistent for all geoinformation producers and users in China. Since 2010, a series of geoinformation projects have been carried out under the leadership of the Ministry of Land and Resources (MLR), including (1) Integration, update and maintenance of geoinformation databases; (2) Standards research on clusterization and industrialization of information services; (3) Platform construction of geological data sharing; (4) Construction of key borehole databases; (5) Product development of information services. "Nine-System" of the basic framework has been proposed for the development and improvement of the geospatial data infrastructure, which are focused on the construction of the cluster organization, cluster service, convergence

  20. Geospatial data infrastructure: The development of metadata for geo-information in China

    International Nuclear Information System (INIS)

    Xu, Baiquan; Yan, Shiqiang; Wang, Qianju; Lian, Jian; Wu, Xiaoping; Ding, Keyong

    2014-01-01

    Stores of geoscience records are in constant flux. These stores are continually added to by new information, ideas and data, which are frequently revised. The geoscience record is in restrained by human thought and technology for handling information. Conventional methods strive, with limited success, to maintain geoscience records which are readily susceptible and renewable. The information system must adapt to the diversity of ideas and data in geoscience and their changes through time. In China, more than 400,000 types of important geological data are collected and produced in geological work during the last two decades, including oil, natural gas and marine data, mine exploration, geophysical, geochemical, remote sensing and important local geological survey and research reports. Numerous geospatial databases are formed and stored in National Geological Archives (NGA) with available formats of MapGIS, ArcGIS, ArcINFO, Metalfile, Raster, SQL Server, Access and JPEG. But there is no effective way to warrant that the quality of information is adequate in theory and practice for decision making. The need for fast, reliable, accurate and up-to-date information by providing the Geographic Information System (GIS) communities are becoming insistent for all geoinformation producers and users in China. Since 2010, a series of geoinformation projects have been carried out under the leadership of the Ministry of Land and Resources (MLR), including (1) Integration, update and maintenance of geoinformation databases; (2) Standards research on clusterization and industrialization of information services; (3) Platform construction of geological data sharing; (4) Construction of key borehole databases; (5) Product development of information services. ''Nine-System'' of the basic framework has been proposed for the development and improvement of the geospatial data infrastructure, which are focused on the construction of the cluster organization, cluster

  1. E-Catalogue “Knowledge Management Practices in Nuclear Organizations”

    International Nuclear Information System (INIS)

    Sheveleva, S.; Pasztory, Z.

    2014-01-01

    The objectives of NKM E-Catalogue: Many nuclear organizations from IAEA Member States have considerable experiences and excellent achievements in the development of Knowledge Management Systems. Depending on organization’s strategy and type of business, they choose various methods and tools of knowledge management for realizing their aims. This catalogue will be available to all Member States interested in learning about collected knowledge management practices in order to enhance their own knowledge management programmes

  2. A geospatial database model for the management of remote sensing datasets at multiple spectral, spatial, and temporal scales

    Science.gov (United States)

    Ifimov, Gabriela; Pigeau, Grace; Arroyo-Mora, J. Pablo; Soffer, Raymond; Leblanc, George

    2017-10-01

    In this study the development and implementation of a geospatial database model for the management of multiscale datasets encompassing airborne imagery and associated metadata is presented. To develop the multi-source geospatial database we have used a Relational Database Management System (RDBMS) on a Structure Query Language (SQL) server which was then integrated into ArcGIS and implemented as a geodatabase. The acquired datasets were compiled, standardized, and integrated into the RDBMS, where logical associations between different types of information were linked (e.g. location, date, and instrument). Airborne data, at different processing levels (digital numbers through geocorrected reflectance), were implemented in the geospatial database where the datasets are linked spatially and temporally. An example dataset consisting of airborne hyperspectral imagery, collected for inter and intra-annual vegetation characterization and detection of potential hydrocarbon seepage events over pipeline areas, is presented. Our work provides a model for the management of airborne imagery, which is a challenging aspect of data management in remote sensing, especially when large volumes of data are collected.

  3. Issues on Building Kazakhstan Geospatial Portal to Implement E-Government

    Science.gov (United States)

    Sagadiyev, K.; Kang, H. K.; Li, K. J.

    2016-06-01

    A main issue in developing e-government is about how to integrate and organize many complicated processes and different stakeholders. Interestingly geospatial information provides an efficient framework to integrate and organized them. In particular, it is very useful to integrate the process of land management in e-government with geospatial information framework, since most of land management tasks are related with geospatial properties. In this paper, we present a use-case on the e-government project in Kazakhstan for land management. We develop a geoportal to connect many tasks and different users via geospatial information framework. This geoportal is based on open source geospatial software including GeoServer, PostGIS, and OpenLayers. With this geoportal, we expect three achievements as follows. First we establish a transparent governmental process, which is one of main goal of e-government. Every stakeholder monitors what is happening in land management process. Second, we can significantly reduce the time and efforts in the government process. For example, a grant procedure for a building construction has taken more than one year with more than 50 steps. It is expected that this procedure would be reduced to 2 weeks by the geoportal framework. Third we provide a collaborative environment between different governmental structures via the geoportal, while many conflicts and mismatches have been a critical issue of governmental administration processes.

  4. ISSUES ON BUILDING KAZAKHSTAN GEOSPATIAL PORTAL TO IMPLEMENT E-GOVERNMENT

    Directory of Open Access Journals (Sweden)

    K. Sagadiyev

    2016-06-01

    Full Text Available A main issue in developing e-government is about how to integrate and organize many complicated processes and different stakeholders. Interestingly geospatial information provides an efficient framework to integrate and organized them. In particular, it is very useful to integrate the process of land management in e-government with geospatial information framework, since most of land management tasks are related with geospatial properties. In this paper, we present a use-case on the e-government project in Kazakhstan for land management. We develop a geoportal to connect many tasks and different users via geospatial information framework. This geoportal is based on open source geospatial software including GeoServer, PostGIS, and OpenLayers. With this geoportal, we expect three achievements as follows. First we establish a transparent governmental process, which is one of main goal of e-government. Every stakeholder monitors what is happening in land management process. Second, we can significantly reduce the time and efforts in the government process. For example, a grant procedure for a building construction has taken more than one year with more than 50 steps. It is expected that this procedure would be reduced to 2 weeks by the geoportal framework. Third we provide a collaborative environment between different governmental structures via the geoportal, while many conflicts and mismatches have been a critical issue of governmental administration processes.

  5. A redesign of OGC Symbology Encoding standard for sharing cartography

    Directory of Open Access Journals (Sweden)

    Erwan Bocher

    2018-01-01

    Full Text Available Despite most Spatial Data Infrastructures offering service-based visualization of geospatial data, requirements are often at a very basic level leading to poor quality of maps. This is a general observation for any geospatial architecture as soon as open standards as those of the Open Geospatial Consortium (OGC are applied. To improve the situation, this paper does focus on improvements at the portrayal interoperability side by considering standardization aspects. We propose two major redesign recommendations. First to consolidate the cartographic theory at the core of the OGC Symbology Encoding standard. Secondly to build the standard in a modular way so as to be ready to be extended with upcoming future cartographic requirements. Thus, we start by defining portrayal interoperability by means of typical-use cases that frame the concept of sharing cartography. Then we bring to light the strengths and limits of the relevant open standards to consider in this context. Finally we propose a set of recommendations to overcome the limits so as to make these use cases a true reality. Even if the definition of a cartographic-oriented standard is not able to act as a complete cartographic design framework by itself, we argue that pushing forward the standardization work dedicated to cartography is a way to share and disseminate good practices and finally to improve the quality of the visualizations.

  6. An MR-compatible neonatal incubator.

    Science.gov (United States)

    Paley, M N J; Hart, A R; Lait, M; Griffiths, P D

    2012-07-01

    To develop a neonatal MR-compatible incubator for transporting babies between a neonatal intensive care unit and an MRI unit that is within the same hospital but geographically separate. The system was strapped to a standard MR-compatible patient trolley, which provides space for resuscitation outside the incubator. A constant-temperature exothermic heat pad was used to maintain temperature together with a logging fluoro-optic temperature monitor and alarm system. The system has been designed to accommodate standard knee-sized coils from the major MR manufacturers. The original incubator was constructed from carbon fibre, but this required modification to prevent radiofrequency shading artefacts due to the conducting properties of the carbon fibre. A high-tensile polyester material was used, which combined light weight with high impact strength. The system could be moved onto the patient bed with the coils and infant in place by one technologist. Studies in eight neonatal patients produced high quality 1.5 T MR images with low motion artefacts. The incubator should also be compatible with imaging in 3 T MR systems, although further work is required to establish this. Images were acquired using both rapid and high-resolution sequences, including three-dimensional volumes, proton spectra and diffusion weighting. The incubator provides a safe, quiet environment for neonates during transport and imaging, at low cost.

  7. Online Resources to Support Professional Development for Managing and Preserving Geospatial Data

    Science.gov (United States)

    Downs, R. R.; Chen, R. S.

    2013-12-01

    tutorials, primers, guides, and online learning modules. The site enables users to find and access standards, real-world examples, and websites of other resources about geospatial data management. Quick links to lists of resources are available for data managers, system developers, and researchers. New resources are featured regularly to highlight current developments in practice and research. A user-centered approach was taken to design and develop the site iteratively, based on a survey of the expectations and needs of community members who have an interest in the management and preservation of geospatial data. Formative and summative evaluation activities have informed design, content, and feature enhancements to enable users to use the website efficiently and effectively. Continuing management and evaluation of the website keeps the content and the infrastructure current with evolving research, practices, and technology. The design, development, evaluation, and use of the website are described along with selected resources and activities that support education and professional development for the management, preservation, and stewardship of geospatial data.

  8. WFCatalog: A catalogue for seismological waveform data

    Science.gov (United States)

    Trani, Luca; Koymans, Mathijs; Atkinson, Malcolm; Sleeman, Reinoud; Filgueira, Rosa

    2017-09-01

    This paper reports advances in seismic waveform description and discovery leading to a new seismological service and presents the key steps in its design, implementation and adoption. This service, named WFCatalog, which stands for waveform catalogue, accommodates features of seismological waveform data. Therefore, it meets the need for seismologists to be able to select waveform data based on seismic waveform features as well as sensor geolocations and temporal specifications. We describe the collaborative design methods and the technical solution showing the central role of seismic feature catalogues in framing the technical and operational delivery of the new service. Also, we provide an overview of the complex environment wherein this endeavour is scoped and the related challenges discussed. As multi-disciplinary, multi-organisational and global collaboration is necessary to address today's challenges, canonical representations can provide a focus for collaboration and conceptual tools for agreeing directions. Such collaborations can be fostered and formalised by rallying intellectual effort into the design of novel scientific catalogues and the services that support them. This work offers an example of the benefits generated by involving cross-disciplinary skills (e.g. data and domain expertise) from the early stages of design, and by sustaining the engagement with the target community throughout the delivery and deployment process.

  9. Planck 2013 results. XXVIII. The Planck Catalogue of Compact Sources

    CERN Document Server

    Ade, P.A.R.; Armitage-Caplan, C.; Arnaud, M.; Ashdown, M.; Atrio-Barandela, F.; Aumont, J.; Baccigalupi, C.; Banday, A.J.; Barreiro, R.B.; Bartlett, J.G.; Battaner, E.; Benabed, K.; Benoit, A.; Benoit-Levy, A.; Bernard, J.P.; Bersanelli, M.; Bielewicz, P.; Bobin, J.; Bock, J.J.; Bonaldi, A.; Bonavera, L.; Bond, J.R.; Borrill, J.; Bouchet, F.R.; Bridges, M.; Bucher, M.; Burigana, C.; Butler, R.C.; Cardoso, J.F.; Carvalho, P.; Catalano, A.; Challinor, A.; Chamballu, A.; Chen, X.; Chiang, L.Y.; Chiang, H.C.; Christensen, P.R.; Church, S.; Clemens, M.; Clements, D.L.; Colombi, S.; Colombo, L.P.L.; Couchot, F.; Coulais, A.; Crill, B.P.; Curto, A.; Cuttaia, F.; Danese, L.; Davies, R.D.; Davis, R.J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Delouis, J.M.; Desert, F.X.; Dickinson, C.; Diego, J.M.; Dole, H.; Donzelli, S.; Dore, O.; Douspis, M.; Dupac, X.; Efstathiou, G.; Ensslin, T.A.; Eriksen, H.K.; Finelli, F.; Forni, O.; Frailis, M.; Franceschi, E.; Galeotta, S.; Ganga, K.; Giard, M.; Giardino, G.; Giraud-Heraud, Y.; Gonzalez-Nuevo, J.; Gorski, K.M.; Gratton, S.; Gregorio, A.; Gruppuso, A.; Hansen, F.K.; Hanson, D.; Harrison, D.; Henrot-Versille, S.; Hernandez-Monteagudo, C.; Herranz, D.; Hildebrandt, S.R.; Hivon, E.; Hobson, M.; Holmes, W.A.; Hornstrup, A.; Hovest, W.; Huffenberger, K.M.; Jaffe, T.R.; Jaffe, A.H.; Jones, W.C.; Juvela, M.; Keihanen, E.; Keskitalo, R.; Kisner, T.S.; Kneissl, R.; Knoche, J.; Knox, L.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lahteenmaki, A.; Lamarre, J.M.; Lasenby, A.; Laureijs, R.J.; Lawrence, C.R.; Leahy, J.P.; Leonardi, R.; Leon-Tavares, J.; Leroy, C.; Lesgourgues, J.; Liguori, M.; Lilje, P.B.; Linden-Vornle, M.; Lopez-Caniego, M.; Lubin, P.M.; Macias-Perez, J.F.; Maffei, B.; Maino, D.; Mandolesi, N.; Maris, M.; Marshall, D.J.; Martin, P.G.; Martinez-Gonzalez, E.; Masi, S.; Matarrese, S.; Matthai, F.; Mazzotta, P.; McGehee, P.; Meinhold, P.R.; Melchiorri, A.; Mendes, L.; Mennella, A.; Migliaccio, M.; Mitra, S.; Miville-Deschenes, M.A.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Munshi, D.; Naselsky, P.; Nati, F.; Natoli, P.; Negrello, M.; Netterfield, C.B.; Norgaard-Nielsen, H.U.; Noviello, F.; Novikov, D.; Novikov, I.; O'Dwyer, I.J.; Osborne, S.; Oxborrow, C.A.; Paci, F.; Pagano, L.; Pajot, F.; Paladini, R.; Paoletti, D.; Partridge, B.; Pasian, F.; Patanchon, G.; Pearson, T.J.; Perdereau, O.; Perotto, L.; Perrotta, F.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Ponthieu, N.; Popa, L.; Poutanen, T.; Pratt, G.W.; Prezeau, G.; Prunet, S.; Puget, J.L.; Rachen, J.P.; Reach, W.T.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Ricciardi, S.; Riller, T.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Roudier, G.; Rowan-Robinson, M.; Rubino-Martin, J.A.; Rusholme, B.; Sandri, M.; Santos, D.; Savini, G.; Schammel, M.P.; Scott, D.; Seiffert, M.D.; Shellard, E.P.S.; Spencer, L.D.; Starck, J.L.; Stolyarov, V.; Stompor, R.; Sudiwala, R.; Sunyaev, R.; Sureau, F.; Sutton, D.; Suur-Uski, A.S.; Sygnet, J.F.; Tauber, J.A.; Tavagnacco, D.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tristram, M.; Tucci, M.; Tuovinen, J.; Turler, M.; Umana, G.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Varis, J.; Vielva, P.; Villa, F.; Vittorio, N.; Wade, L.A.; Walter, B.; Wandelt, B.D.; Yvon, D.; Zacchei, A.; Zonca, A.

    2014-01-01

    The Planck Catalogue of Compact Sources (PCCS) is the catalogue of sources detected in the first 15 months of Planck operations, the "nominal" mission. It consists of nine single-frequency catalogues of compact sources, both Galactic and extragalactic, detected over the entire sky. The PCCS covers the frequency range 30--857\\,GHz with higher sensitivity (it is 90% complete at 180 mJy in the best channel) and better angular resolution (from ~33' to ~5') than previous all-sky surveys in this frequency band. By construction its reliability is >80% and more than 65% of the sources have been detected at least in two contiguous Planck channels. In this paper we present the construction and validation of the PCCS, its contents and its statistical characterization.

  10. Application of remote handling compatibility on ITER plant

    International Nuclear Information System (INIS)

    Sanders, S.; Rolfe, A.; Mills, S.F.; Tesini, A.

    2011-01-01

    The ITER plant will require fully remote maintenance during its operational life. For this to be effective, safe and efficient the plant will have to be developed in accordance with remote handling (RH) compatibility requirements. A system for ensuring RH compatibility on plant designed for Tokamaks was successfully developed and applied, inter alia, by the authors when working at the JET project. The experience gained in assuring RH compatibility of plant at JET is now being applied to RH relevant ITER plant. The methodologies required to ensure RH compatibility of plant include the standardization of common plant items, standardization of RH features, availability of common guidance on RH best practice and a protocol for design and interface review and approval. The protocol in use at ITER is covered by the ITER Remote Maintenance Management System (IRMMS) defines the processes and utilization of management controls including Plant Definition Forms (PDF), Task Definition Forms (TDFs) and RH Compatibility Assessment Forms (RHCA) and the ITER RH Code of Practice. This paper will describe specific examples where the authors have applied the methodology proven at JET to ensure remote handling compatibility on ITER plant. Examples studied are: ·ELM coils (to be installed in-vessel behind the Blanket Modules) - handling both in-vessel, in Casks and at the Hot Cell as well as fully remote installation and connection (mechanical and electrical) in-vessel. ·Neutral beam systems (in-vessel and in the NB Cell) - beam sources, cesium oven, beam line components (accessed in the NB Cell) and Duct Liner (remotely replaced from in-vessel). ·Divertor (in-vessel) - cooling pipe work and remotely operated electrical connector. The RH compatibility process can significantly affect plant design. This paper should therefore be of interest to all parties who develop ITER plant designs.

  11. Exploring best cataloguing rules in the 21st century: Changes from ...

    African Journals Online (AJOL)

    In this digital era, the need to embrace change is inevitable. The authors described fundamental changes that were necessary to move cataloguing practice to the next level. Some of these changes include but not limited to: cataloguing working tools, changes in information resources, vocabulary, main entry points and ...

  12. GIBS Geospatial Data Abstraction Library (GDAL)

    Data.gov (United States)

    National Aeronautics and Space Administration — GDAL is an open source translator library for raster geospatial data formats that presents a single abstract data model to the calling application for all supported...

  13. Searching the databases: a quick look at Amazon and two other online catalogues.

    Science.gov (United States)

    Potts, Hilary

    2003-01-01

    The Amazon Online Catalogue was compared with the Library of Congress Catalogue and the British Library Catalogue, both also available online, by searching on both neutral (Gay, Lesbian, Homosexual) and pejorative (Perversion, Sex Crime) subject terms, and also by searches using Boolean logic in an attempt to identify Lesbian Fiction items and religion-based anti-gay material. Amazon was much more likely to be the first port of call for non-academic enquiries. Although excluding much material necessary for academic research, it carried more information about the individual books and less historical homophobic baggage in its terminology than the great national catalogues. Its back catalogue of second-hand books outnumbered those in print. Current attitudes may partially be gauged by the relative numbers of titles published under each heading--e.g., there may be an inverse relationship between concern about child sex abuse and homophobia, more noticeable in U.S. because of the activities of the religious right.

  14. Geospatial Information System Capability Maturity Models

    Science.gov (United States)

    2017-06-01

    To explore how State departments of transportation (DOTs) evaluate geospatial tool applications and services within their own agencies, particularly their experiences using capability maturity models (CMMs) such as the Urban and Regional Information ...

  15. Enlightening music: the catalogue and digitisation project of Verdi’s archive at the “Archivio Storico Ricordi” in Milan

    Directory of Open Access Journals (Sweden)

    Mathias Balbi

    2014-05-01

    Full Text Available Our contribution concentrates on the processes of cataloguing and digitising of hand–written letters and business documents belonging to the ''Ricordi Historical Archive'' in Milan, which is at present housed in the Braidense National Library; the archive possesses documents referring to the history of the publishing company covering the two centuries of its activities. Starting from the ledgers compiled by Giovanni Ricordi, which cover a period from 1814 up to the middle of the 19th century to the documents relative to the school of engraving, the archive also houses a large number of editorial catalogues, published by Ricordi and by other publishing houses which it took over, as well as the original contracts stipulated with composers and librettists from Giuseppe Verdi and Giacomo Puccini to Arrigo Boito. The collection of about 15,000 hand–written letters covers the period from the beginning of the 19th century to the end of the 20th century and includes documents sent to Ricordi from writers, singers authors, and composers such as Giuseppe Verdi, Giacomo Puccini and Ottorino Respighi and librettists Luigi Illica, Giuseppe Giacosa and Arrigo Boito. The second part of the correspondence is made up letter–books and business correspondence which were kept intact from 1888 to 1962, except for the period from 1944 to 1953. Since 2006, work has been carried out to catalogue and digitise the Ricordi Archive based on the standards set out by the National Library Service (Servizio Bibliotecario Nazionale – SBN which is head of the Central Institute for the Single Catalogue (ICCU. The structure and use of the archive and the external transmission of its contents are safeguarded and enhanced thanks to the combined work of cataloguing and digitisation of this heritage: the preliminary cataloguing of the documents uses a simplified model of the scheme adopted in the description of the manuscript according to sw Manus, elaborated by ICCU: the

  16. Roma-BZCAT: a multifrequency catalogue of blazars

    Science.gov (United States)

    Massaro, E.; Giommi, P.; Leto, C.; Marchegiani, P.; Maselli, A.; Perri, M.; Piranomonte, S.; Sclavi, S.

    2009-02-01

    We present a new catalogue of blazars based on multifrequency surveys and on an extensive review of the literature. Blazars are classified as BL Lacertae objects, as flat spectrum radio quasars or as blazars of uncertain/transitional type. Each object is identified by a root name, coded as BZB, BZQ and BZU for these three subclasses respectively, and by its coordinates. This catalogue is being built as a tool useful for the identification of the extragalactic sources that will be detected by present and future experiments for X and gamma-ray astronomy, like Swift, AGILE, Fermi-GLAST and Simbol-X. An electronic version is available from the ASI Science Data Center web site at http://www.asdc.asi.it/bzcat.

  17. LU60645GT and MA132843GT Catalogues of Lunar and Martian Impact Craters Developed Using a Crater Shape-based Interpolation Crater Detection Algorithm for Topography Data

    Science.gov (United States)

    Salamuniccar, Goran; Loncaric, Sven; Mazarico, Erwan Matias

    2012-01-01

    For Mars, 57,633 craters from the manually assembled catalogues and 72,668 additional craters identified using several crater detection algorithms (CDAs) have been merged into the MA130301GT catalogue. By contrast, for the Moon the most complete previous catalogue contains only 14,923 craters. Two recent missions provided higher-quality digital elevation maps (DEMs): SELENE (in 1/16° resolution) and Lunar Reconnaissance Orbiter (we used up to 1/512°). This was the main motivation for work on the new Crater Shape-based interpolation module, which improves previous CDA as follows: (1) it decreases the number of false-detections for the required number of true detections; (2) it improves detection capabilities for very small craters; and (3) it provides more accurate automated measurements of craters' properties. The results are: (1) LU60645GT, which is currently the most complete (up to D>=8 km) catalogue of Lunar craters; and (2) MA132843GT catalogue of Martian craters complete up to D>=2 km, which is the extension of the previous MA130301GT catalogue. As previously achieved for Mars, LU60645GT provides all properties that were provided by the previous Lunar catalogues, plus: (1) correlation between morphological descriptors from used catalogues; (2) correlation between manually assigned attributes and automated measurements; (3) average errors and their standard deviations for manually and automatically assigned attributes such as position coordinates, diameter, depth/diameter ratio, etc; and (4) a review of positional accuracy of used datasets. Additionally, surface dating could potentially be improved with the exhaustiveness of this new catalogue. The accompanying results are: (1) the possibility of comparing a large number of Lunar and Martian craters, of e.g. depth/diameter ratio and 2D profiles; (2) utilisation of a method for re-projection of datasets and catalogues, which is very useful for craters that are very close to poles; and (3) the extension of the

  18. Open geospatial infrastructure for data management and analytics in interdisciplinary research

    DEFF Research Database (Denmark)

    Jeppesen, Jacob Høxbroe; Ebeid, Emad Samuel Malki; Jacobsen, Rune Hylsberg

    2018-01-01

    , and information and communications technology needed to promote the implementation of precision agriculture is limited by proprietary integrations and non-standardized data formats and connections. In this paper, an open geospatial data infrastructure is presented, based on standards defined by the Open...... software, and was complemented by open data from governmental offices along with ESA satellite imagery. Four use cases are presented, covering analysis of nearly 50 000 crop fields and providing seamless interaction with an emulated machine terminal. They act to showcase both for how the infrastructure......The terms Internet of Things and Big Data are currently subject to much attention, though the specific impact of these terms in our practical lives are difficult to apprehend. Data-driven approaches do lead to new possibilities, and significant improvements within a broad range of domains can...

  19. Clay club catalogue of characteristics of argillaceous rocks

    International Nuclear Information System (INIS)

    2005-01-01

    The OECD/NEA Working Group on the Characterisation, the Understanding and the Performance of Argillaceous Rocks as Repository Host Formations, namely the Clay Club, examines the various argillaceous rocks that are being considered for the deep geological disposal of radioactive waste, i.e. from plastic, soft, poorly indurated clays to brittle, hard mud-stones or shales. The Clay Club considered it necessary and timely to provide a catalogue to gather in a structured way the key geo-scientific characteristics of the various argillaceous formations that are - or were - studied in NEA member countries with regard to radioactive waste disposal. The present catalogue represents the outcomes of this Clay Club initiative. (author)

  20. Towards Geo-spatial Information Science in Big Data Era

    Directory of Open Access Journals (Sweden)

    LI Deren

    2016-04-01

    Full Text Available Since the 1990s, with the advent of worldwide information revolution and the development of internet, geospatial information science have also come of age, which pushed forward the building of digital Earth and cyber city. As we entered the 21st century, with the development and integration of global information technology and industrialization, internet of things and cloud computing came into being, human society enters into the big data era. This article covers the key features (ubiquitous, multi-dimension and dynamics, internet+networking, full automation and real-time, from sensing to recognition, crowdsourcing and VGI, and service-oriented of geospatial information science in the big data era and addresses the key technical issues (non-linear four dimensional Earth reference frame system, space based enhanced GNSS, space-air and land unified network communication techniques, on board processing techniques for multi-sources image data, smart interface service techniques for space-borne information, space based resource scheduling and network security, design and developing of a payloads based multi-functional satellite platform. That needs to be resolved to provide a new definition of geospatial information science in big data era. Based on the discussion in this paper, the author finally proposes a new definition of geospatial information science (geomatics, i.e. Geomatics is a multiple discipline science and technology which, using a systematic approach, integrates all the means for spatio-temporal data acquisition, information extraction, networked management, knowledge discovering, spatial sensing and recognition, as well as intelligent location based services of any physical objects and human activities around the earth and its environment. Starting from this new definition, geospatial information science will get much more chances and find much more tasks in big data era for generation of smart earth and smart city . Our profession

  1. European wind turbine catalogue

    International Nuclear Information System (INIS)

    1994-01-01

    The THERMIE European Community programme is designed to promote the greater use of European technology and this catalogue contributes to the fulfillment of this aim by dissemination of information on 50 wind turbines from 30 manufacturers. These turbines are produced in Europe and are commercially available. The manufacturers presented produce and sell grid-connected turbines which have been officially approved in countries where this approval is acquired, however some of the wind turbines included in the catalogue have not been regarded as fully commercially available at the time of going to print. The entries, which are illustrated by colour photographs, give company profiles, concept descriptions, measured power curves, prices, and information on design and dimension, safety systems, stage of development, special characteristics, annual energy production, and noise pollution. Lists are given of wind turbine manufacturers and agents and of consultants and developers in the wind energy sector. Exchange rates used in the conversion of the prices of wind turbines are also given. Information can be found on the OPET network (organizations recognised by the European Commission as an Organization for the Promotion of Energy Technologies (OPET)). An article describes the development of the wind power industry during the last 10-15 years and another article on certification aims to give an overview of the most well-known and acknowledged type approvals currently issued in Europe. (AB)

  2. Student Focused Geospatial Curriculum Initiatives: Internships and Certificate Programs at NCCU

    Science.gov (United States)

    Vlahovic, G.; Malhotra, R.

    2009-12-01

    This paper reports recent efforts by the Department of Environmental, Earth and Geospatial Sciences faculty at North Carolina Central University (NCCU) to develop a leading geospatial sciences program that will be considered a model for other Historically Black College/University (HBCU) peers nationally. NCCU was established in 1909 and is the nation’s first state supported public liberal arts college funded for African Americans. In the most recent annual ranking of America’s best black colleges by the US News and World Report (Best Colleges 2010), NCCU was ranked 10th in the nation. As one of only two HBCUs in the southeast offering an undergraduate degree in Geography (McKee, J.O. and C. V. Dixon. Geography in Historically Black Colleges/ Universities in the Southeast, in The Role of the South in Making of American Geography: Centennial of the AAG, 2004), NCCU is uniquely positioned to positively affect talent and diversity of the geospatial discipline in the future. Therefore, successful creation of research and internship pathways for NCCU students has national implications because it will increase the number of minority students joining the workforce and applying to PhD programs. Several related efforts will be described, including research and internship projects with Fugro EarthData Inc., Center for Remote Sensing and Mapping Science at the University of Georgia, Center for Earthquake Research and Information at the University of Memphis and the City of Durham. The authors will also outline requirements and recent successes of ASPRS Provisional Certification Program, developed and pioneered as collaborative effort between ASPRS and NCCU. This certificate program allows graduating students majoring in geospatial technologies and allied fields to become provisionally certified by passing peer-review and taking the certification exam. At NCCU, projects and certification are conducted under the aegis of the Geospatial Research, Innovative Teaching and

  3. Electric charge quantisation from gauge invariance of a Lagrangian: a catalogue of baryon number violating scalar interactions

    International Nuclear Information System (INIS)

    Bowes, J.P.; Foot, R.; Volkas, R.R.

    1997-01-01

    In gauge theories like the standard model, the electric charges of the fermions can be heavily constrained from the classical structure of the theory and from the cancellation of anomalies. There is however mounting evidence suggesting that these anomaly constraints are not as well motivated as the classical constraints. In light of this, possible modifications of the minimal standard model are discussed which will give a complete electric charge quantisation from classical constraints alone. Because these modifications to the Standard Model involve the consideration of baryon number violating scalar interactions, a complete catalogue of the simplest ways to modify the Standard Model is presented so as to introduce explicit baryon number violation. 9 refs., 7 figs

  4. Electric charge quantisation from gauge invariance of a Lagrangian: a catalogue of baryon number violating scalar interactions

    Energy Technology Data Exchange (ETDEWEB)

    Bowes, J.P.; Foot, R.; Volkas, R.R.

    1997-06-01

    In gauge theories like the standard model, the electric charges of the fermions can be heavily constrained from the classical structure of the theory and from the cancellation of anomalies. There is however mounting evidence suggesting that these anomaly constraints are not as well motivated as the classical constraints. In light of this, possible modifications of the minimal standard model are discussed which will give a complete electric charge quantisation from classical constraints alone. Because these modifications to the Standard Model involve the consideration of baryon number violating scalar interactions, a complete catalogue of the simplest ways to modify the Standard Model is presented so as to introduce explicit baryon number violation. 9 refs., 7 figs.

  5. Catalogue of tide gauges in the Pacific

    National Research Council Canada - National Science Library

    Ridgway, N. M

    1984-01-01

    Although this catalogue is primarily intended to provide a list of sources for tidal data which can be used in postevent studies of tsunamis, it may also be useful in other branches of oceanographic...

  6. Catalogue of response spectra for unfolding in situ gamma-ray pulse-height distributions

    International Nuclear Information System (INIS)

    Dymke, N.

    1982-01-01

    To unfold in situ gamma-ray pulse-height distributions by means of a response matrix technique, the matrix must be in keeping with the measurement geometry, detector size, and energy range to be covered by the measurements. A methodology has been described for determination of standard gamma-ray spectra needed in deriving response matrices and a spectrum catalogue compiled containing graphs and data for the 0-3 MeV (4 x 4 in. NaI(Tl)) and 0-8 MeV (1.5 x 1.5 in. NaI(Tl)) ranges. (author)

  7. New Techniques for Deep Learning with Geospatial Data using TensorFlow, Earth Engine, and Google Cloud Platform

    Science.gov (United States)

    Hancher, M.

    2017-12-01

    Recent years have seen promising results from many research teams applying deep learning techniques to geospatial data processing. In that same timeframe, TensorFlow has emerged as the most popular framework for deep learning in general, and Google has assembled petabytes of Earth observation data from a wide variety of sources and made them available in analysis-ready form in the cloud through Google Earth Engine. Nevertheless, developing and applying deep learning to geospatial data at scale has been somewhat cumbersome to date. We present a new set of tools and techniques that simplify this process. Our approach combines the strengths of several underlying tools: TensorFlow for its expressive deep learning framework; Earth Engine for data management, preprocessing, postprocessing, and visualization; and other tools in Google Cloud Platform to train TensorFlow models at scale, perform additional custom parallel data processing, and drive the entire process from a single familiar Python development environment. These tools can be used to easily apply standard deep neural networks, convolutional neural networks, and other custom model architectures to a variety of geospatial data structures. We discuss our experiences applying these and related tools to a range of machine learning problems, including classic problems like cloud detection, building detection, land cover classification, as well as more novel problems like illegal fishing detection. Our improved tools will make it easier for geospatial data scientists to apply modern deep learning techniques to their own problems, and will also make it easier for machine learning researchers to advance the state of the art of those techniques.

  8. VizieR Online Data Catalog: Catalogue of Stellar Spectral Classifications (Skiff, 2010)

    Science.gov (United States)

    Skiff, B. A.

    2009-02-01

    Morgan & Abt 'MKA' paper (1972AJ.....77...35M) not appearing in the two later lists are added. Keenan made continual adjustments to the standards lists up to the time of his death. Thus the late-type standards comprise those marked as high-weight standards in the 1989 Perkins catalogue (1989ApJS...71..245K = III/150); the revised S-type standards in collaboration with Boeshaar (1980ApJS...43..379K); plus the carbon standards and class IIIb 'clump giants' in collaboration with Barnbaum (1996ApJS..105..419B and 1999ApJ...518..859K). In addition, I have made use of the final types by Keenan up to January 2000 shown at the Ohio State Web site (http://www.astronomy.ohio-state.edu/MKCool), accessed in autumn 2003. Though the present file contains all the stars in these lists, only those marked as standards are flagged as such. Garrison's list of MK 'anchor points' might also be consulted in this regard (1994mpyp.conf....3G, and http://www.astro.utoronto.ca/~garrison/mkstds.html). The catalogue includes for the first time results from many large-scale objective-prism spectral surveys done at Case, Stockholm, Crimea, Abastumani, and elsewhere. The stars in these surveys were usually identified only on charts or by other indirect means, and have been overlooked heretofore because of the difficulty in recovering the stars. More complete results from these separate publications, including notes and identifications, have been made available to the CDS, and are kept at the Lowell Observatory ftp area (ftp://ftp.lowell.edu/pub/bas/starcats). Not all of these stars are present in SIMBAD. As a 'living catalogue', an attempt will be made to keep up with current literature, and to extend the indexing of citations back in time. (2 data files).

  9. Web mapping system for complex processing and visualization of environmental geospatial datasets

    Science.gov (United States)

    Titov, Alexander; Gordov, Evgeny; Okladnikov, Igor

    2016-04-01

    Environmental geospatial datasets (meteorological observations, modeling and reanalysis results, etc.) are used in numerous research applications. Due to a number of objective reasons such as inherent heterogeneity of environmental datasets, big dataset volume, complexity of data models used, syntactic and semantic differences that complicate creation and use of unified terminology, the development of environmental geodata access, processing and visualization services as well as client applications turns out to be quite a sophisticated task. According to general INSPIRE requirements to data visualization geoportal web applications have to provide such standard functionality as data overview, image navigation, scrolling, scaling and graphical overlay, displaying map legends and corresponding metadata information. It should be noted that modern web mapping systems as integrated geoportal applications are developed based on the SOA and might be considered as complexes of interconnected software tools for working with geospatial data. In the report a complex web mapping system including GIS web client and corresponding OGC services for working with geospatial (NetCDF, PostGIS) dataset archive is presented. There are three basic tiers of the GIS web client in it: 1. Tier of geospatial metadata retrieved from central MySQL repository and represented in JSON format 2. Tier of JavaScript objects implementing methods handling: --- NetCDF metadata --- Task XML object for configuring user calculations, input and output formats --- OGC WMS/WFS cartographical services 3. Graphical user interface (GUI) tier representing JavaScript objects realizing web application business logic Metadata tier consists of a number of JSON objects containing technical information describing geospatial datasets (such as spatio-temporal resolution, meteorological parameters, valid processing methods, etc). The middleware tier of JavaScript objects implementing methods for handling geospatial

  10. Seminar program for postgraduate specialty training in general practice: proposal for a 5-year thematic catalogue

    Directory of Open Access Journals (Sweden)

    Sommer, Susanne

    2017-11-01

    Full Text Available Introduction: In different German regions, seminar programs have been conducted for General practice residents. In each region, selection and teaching of learning content is conducted in a different manner. So far, no structured, standardized curriculum has been implemented nationwide. We have investigated, if the development of a common 5-year program of learning topics is conceivable between the different university departments of General practice in Germany.Method: The seminar program working group of the DEGAM (German College of General Practitioners and Family Physicians has conducted an online survey based on information gathered via preliminary telephone conference (n=7; physicians with postgraduate teaching experience among all German university departments of General Practice and two non-university teaching institutions, identified via the internet. 884 topics were extracted from 14 Seminar programs. The topics were entered in a database, discussed and categorized: Practice management/practice work flow/standardized documentation forms/quality management (n=33 topics, common acute and chronic diseases, including disease management programs (n=29 topics, communication, neurological, psychological and psychiatric consultations (n=24 topics, common medical problems, including eye, ear, nose, throat, skin and pediatric problems (n=99 Topics family physicians general approach, including epidemiology, shared decision making, test of time (n=42 Topics. These topics have been rated for priority and desirable number of teaching-units.Results: A catalogue of 111 topics was designed, encompassing 160 teaching units. There is a suggestion of wide topics collections plus an add-on catalogue.Conclusion: A proposal for a 5-year-thematic catalogue for postgraduate training of general practice residents in Germany has been developed. This newly developed curriculum has the potential to improve knowledge and skills that have not been covered during in

  11. Transport Infrastructure in the Process of Cataloguing Brownfields

    Science.gov (United States)

    Kramářová, Zuzana

    2017-10-01

    To begin with, the identification and follow-up revitalisation of brownfields raises a burning issue in territorial planning as well as in construction engineering. This phenomenon occurs not only in the Czech Republic and Europe, but also world-wide experts conduct its careful investigation. These issues may be divided into several areas. First, it is identifying and cataloguing single territorial localities; next, it means a complex process of locality revitalisation. As a matter of fact, legislative framework represents a separate area, which is actually highly specific in individual countries in accordance with the existing law, norms and regulations (it concerns mainly territorial planning and territory segmentation into appropriate administrative units). Legislative base of the Czech Republic was analysed in an article at WMCAUS in 2016. The solution of individual identification and following cataloguing of brownfields is worked out by Form of Regional Studies within the Legislation of the Czech Republic. Due to huge the scale of issues to be tackled, their content is only loosely defined in regard to Building Act and its implementing regulations, e.g. examining the layout of future construction in the area, locating architecturally or otherwise interesting objects, transport or technical infrastructure management, tourism, socially excluded localities etc. Legislative base does not exist, there is no common method for identifying and cataloguing brownfields. Therefore, individual catalogue lists are subject to customer’s requirements. All the same, the relevant information which the database contains may be always examined. One of them is part about transport infrastructure. The information may be divided into three subareas - information on transport accessibility of the locality, information on the actual infrastructure in the locality and information on the transport accessibility of human resources.

  12. Scalable global grid catalogue for Run3 and beyond

    Science.gov (United States)

    Martinez Pedreira, M.; Grigoras, C.; ALICE Collaboration

    2017-10-01

    The AliEn (ALICE Environment) file catalogue is a global unique namespace providing mapping between a UNIX-like logical name structure and the corresponding physical files distributed over 80 storage elements worldwide. Powerful search tools and hierarchical metadata information are integral parts of the system and are used by the Grid jobs as well as local users to store and access all files on the Grid storage elements. The catalogue has been in production since 2005 and over the past 11 years has grown to more than 2 billion logical file names. The backend is a set of distributed relational databases, ensuring smooth growth and fast access. Due to the anticipated fast future growth, we are looking for ways to enhance the performance and scalability by simplifying the catalogue schema while keeping the functionality intact. We investigated different backend solutions, such as distributed key value stores, as replacement for the relational database. This contribution covers the architectural changes in the system, together with the technology evaluation, benchmark results and conclusions.

  13. Studies on Colombian Cryptogams IV. A catalogue of the Hepaticae of Colombia

    NARCIS (Netherlands)

    Gradstein, S.R.; Hekking, W.H.A.

    1979-01-01

    □ The main purpose of this catalogue is to provide a complete listing of the species of liverworts hitherto known from Colombia and to summarize our present knowledge of species distribution within the country. It was prepared parallel to a catalogue of the mosses (Musci), which is being published

  14. Planck 2015 results: XXVII. The second Planck catalogue of Sunyaev-Zeldovich sources

    DEFF Research Database (Denmark)

    Ade, P. A R; Aghanim, N.; Arnaud, M.

    2016-01-01

    We present the all-sky Planck catalogue of Sunyaev-Zeldovich (SZ) sources detected from the 29 month full-mission data. The catalogue (PSZ2) is the largest SZ-selected sample of galaxy clusters yet produced and the deepest systematic all-sky surveyof galaxy clusters. It contains 1653 detections, ...

  15. MapFactory - Towards a mapping design pattern for big geospatial data

    Science.gov (United States)

    Rautenbach, Victoria; Coetzee, Serena

    2018-05-01

    With big geospatial data emerging, cartographers and geographic information scientists have to find new ways of dealing with the volume, variety, velocity, and veracity (4Vs) of the data. This requires the development of tools that allow processing, filtering, analysing, and visualising of big data through multidisciplinary collaboration. In this paper, we present the MapFactory design pattern that will be used for the creation of different maps according to the (input) design specification for big geospatial data. The design specification is based on elements from ISO19115-1:2014 Geographic information - Metadata - Part 1: Fundamentals that would guide the design and development of the map or set of maps to be produced. The results of the exploratory research suggest that the MapFactory design pattern will help with software reuse and communication. The MapFactory design pattern will aid software developers to build the tools that are required to automate map making with big geospatial data. The resulting maps would assist cartographers and others to make sense of big geospatial data.

  16. Center of Excellence for Geospatial Information Science research plan 2013-18

    Science.gov (United States)

    Usery, E. Lynn

    2013-01-01

    The U.S. Geological Survey Center of Excellence for Geospatial Information Science (CEGIS) was created in 2006 and since that time has provided research primarily in support of The National Map. The presentations and publications of the CEGIS researchers document the research accomplishments that include advances in electronic topographic map design, generalization, data integration, map projections, sea level rise modeling, geospatial semantics, ontology, user-centered design, volunteer geographic information, and parallel and grid computing for geospatial data from The National Map. A research plan spanning 2013–18 has been developed extending the accomplishments of the CEGIS researchers and documenting new research areas that are anticipated to support The National Map of the future. In addition to extending the 2006–12 research areas, the CEGIS research plan for 2013–18 includes new research areas in data models, geospatial semantics, high-performance computing, volunteered geographic information, crowdsourcing, social media, data integration, and multiscale representations to support the Three-Dimensional Elevation Program (3DEP) and The National Map of the future of the U.S. Geological Survey.

  17. Modern Special Collections Cataloguing: A University of London Case Study

    OpenAIRE

    Attar, Karen

    2013-01-01

    Recent years have seen a growing emphasis on modern special collections (in themselves no new phenomenon), with a dichotomy between guidance for detailed cataloguing in Descriptive Cataloging of Rare Materials (Books) (DCRM(B), 2007) and the value of clearing cataloguing backlogs expeditiously. This article describes the De la Mare Family Archive of Walter de la Mare's Printed Oeuvre at Senate House Library, University of London, as an example of a modern author collections in an institutiona...

  18. IAEA Catalogue of Services for Nuclear Infrastructure Development. Rev. 1, April 2014

    International Nuclear Information System (INIS)

    2014-04-01

    This IAEA Catalogue offers a wide range of services to Member States embarking on a new nuclear power programme or expanding an existing one. A new IAEA Catalogue of Services for Nuclear Infrastructure Development helps Member States to identify and request IAEA assistance for national organizations at different stages of the development or expansion of a nuclear power programme. This IAEA Catalogue of Services is presented in two tables. It is based on the IAEA Milestones Approach for nuclear power infrastructure development, documented in 'Milestones in the Development of a National Infrastructure for Nuclear Power' (IAEA Nuclear Energy Series NG-G-3.1). The two tables allow users to identify and select available IAEA services by: i) The three phases of the IAEA Milestones Approach, or ii) Organizations typically involved in the development of a nuclear power programme: the government / Nuclear Energy Programme Implementing Organization (NEPIO), the regulatory body and the owner operator of a nuclear power plant. This Catalogue includes information on the following IAEA services: i) Workshops / Training Courses; ii) Expert Missions / Advisory Services; iii) Review Missions / Peer Reviews; iv) Training tools and networks. The Catalogue lists both existing IAEA services and those being developed for the 19 issues to be addressed in developing a national nuclear infrastructure. Each existing service is linked to a relevant IAEA webpage that either describes a particular service or gives practical examples of the type of assistance that the Agency offers (e.g. workshops or missions). The owners of these webpages can be contacted for more detailed information or to request assistance. This IAEA Catalogue of Services will be updated regularly

  19. Strategizing Teacher Professional Development for Classroom Uses of Geospatial Data and Tools

    Science.gov (United States)

    Zalles, Daniel R.; Manitakos, James

    2016-01-01

    Studying Topography, Orographic Rainfall, and Ecosystems with Geospatial Information Technology (STORE), a 4.5-year National Science Foundation funded project, explored the strategies that stimulate teacher commitment to the project's driving innovation: having students use geospatial information technology (GIT) to learn about weather, climate,…

  20. An application of data mining techniques in designing catalogue for a laundry service

    Directory of Open Access Journals (Sweden)

    Khasanah Annisa Uswatun

    2018-01-01

    Full Text Available Catalogues are the media that companies use to promote their products or services. Since catalogue is one of marketing media, the first essential step before designing product catalogue is determining the market target. Besides, it is also important to put some information that appeal to the target market, such as discount or promos by analysing customer pattern preferences in using services or buying product. This study conduct two data mining technique. The first is clustering analysis to segment customer and the second one is association rule mining to discover an interesting pattern about the services that commonly used by the customer at the same service time. Thus, the results will be used as a recommendation to make an attractive marketing strategy to be put in the service catalogue promo for a laundry in Sleman Yogyakarta. The clustering result showed that the biggest customer segment is university student who come 3 until 5 times in a month on weekends, while the association rule result showed that clothes, shoes, and bed sheet have strong relationship. The catalogue design is presented in the end of the paper.

  1. Geospatial Big Data Handling Theory and Methods: A Review and Research Challenges

    DEFF Research Database (Denmark)

    Li, Songnian; Dragicevic, Suzana; Anton, François

    2016-01-01

    Big data has now become a strong focus of global interest that is increasingly attracting the attention of academia, industry, government and other organizations. Big data can be situated in the disciplinary area of traditional geospatial data handling theory and methods. The increasing volume...... for Photogrammetry and Remote Sensing (ISPRS) Technical Commission II (TC II) revisits the existing geospatial data handling methods and theories to determine if they are still capable of handling emerging geospatial big data. Further, the paper synthesises problems, major issues and challenges with current...... developments as well as recommending what needs to be developed further in the near future....

  2. Towards the Development of a Taxonomy for Visualisation of Streamed Geospatial Data

    Science.gov (United States)

    Sibolla, B. H.; Van Zyl, T.; Coetzee, S.

    2016-06-01

    Geospatial data has very specific characteristics that need to be carefully captured in its visualisation, in order for the user and the viewer to gain knowledge from it. The science of visualisation has gained much traction over the last decade as a response to various visualisation challenges. During the development of an open source based, dynamic two-dimensional visualisation library, that caters for geospatial streaming data, it was found necessary to conduct a review of existing geospatial visualisation taxonomies. The review was done in order to inform the design phase of the library development, such that either an existing taxonomy can be adopted or extended to fit the needs at hand. The major challenge in this case is to develop dynamic two dimensional visualisations that enable human interaction in order to assist the user to understand the data streams that are continuously being updated. This paper reviews the existing geospatial data visualisation taxonomies that have been developed over the years. Based on the review, an adopted taxonomy for visualisation of geospatial streaming data is presented. Example applications of this taxonomy are also provided. The adopted taxonomy will then be used to develop the information model for the visualisation library in a further study.

  3. Studies on compatibility of energetic materials by thermal methods

    Directory of Open Access Journals (Sweden)

    Maria Alice Carvalho Mazzeu

    2010-04-01

    Full Text Available The chemical compatibility of explosives, pyrotechnics and propellants with those materials is studied to evaluate potential hazards when in contact with other materials during production, storage and handling. Compatibility can be studied by several thermal methods as DSC (differential scanning calorimetry, TG (Thermogravimetry, VST (Vacuum stability test and others. The test methods and well defined criteria are the most important elements when a compatibility study is being accomplished. In this paper, the compatibility of two very important high explosives used in ammunition, RDX (Cyclo-1,3,5-trimethylene-2,4,6-trinitramine and HMX (Cyclotetramethylene tetranitramine was studied with the materials: fluoroelastomer (Viton and powdered aluminum (Al, using DSC and VST methods. The criteria to judge the compatibility between materials is based on a standardization agreement (STANAG 4147, 2001, and the final conclusion is that explosives and this materials are compatible, but in DSC it was observed that the peak of decomposition temperature of the admixture of RDX with Al decreased in 3º C and another peak appeared after the decomposition peak.

  4. The compiled catalogue of galaxies in machine-readable form and its statistical investigation

    International Nuclear Information System (INIS)

    Kogoshvili, N.G.

    1982-01-01

    The compilation of a machine-readable catalogue of relatively bright galaxies was undertaken in Abastumani Astrophysical Observatory in order to facilitate the statistical analysis of a large observational material on galaxies from the Palomar Sky Survey. In compiling the catalogue of galaxies the following problems were considered: the collection of existing information for each galaxy; a critical approach to data aimed at the selection of the most important features of the galaxies; the recording of data in computer-readable form; and the permanent updating of the catalogue. (Auth.)

  5. Planck 2013 results. XXXII. The updated Planck catalogue of Sunyaev-Zeldovich sources

    Science.gov (United States)

    Planck Collaboration; Ade, P. A. R.; Aghanim, N.; Armitage-Caplan, C.; Arnaud, M.; Ashdown, M.; Atrio-Barandela, F.; Aumont, J.; Aussel, H.; Baccigalupi, C.; Banday, A. J.; Barreiro, R. B.; Barrena, R.; Bartelmann, M.; Bartlett, J. G.; Battaner, E.; Benabed, K.; Benoît, A.; Benoit-Lévy, A.; Bernard, J.-P.; Bersanelli, M.; Bielewicz, P.; Bikmaev, I.; Bobin, J.; Bock, J. J.; Böhringer, H.; Bonaldi, A.; Bond, J. R.; Borrill, J.; Bouchet, F. R.; Bridges, M.; Bucher, M.; Burenin, R.; Burigana, C.; Butler, R. C.; Cardoso, J.-F.; Carvalho, P.; Catalano, A.; Challinor, A.; Chamballu, A.; Chary, R.-R.; Chen, X.; Chiang, H. C.; Chiang, L.-Y.; Chon, G.; Christensen, P. R.; Churazov, E.; Church, S.; Clements, D. L.; Colombi, S.; Colombo, L. P. L.; Comis, B.; Couchot, F.; Coulais, A.; Crill, B. P.; Curto, A.; Cuttaia, F.; Da Silva, A.; Dahle, H.; Danese, L.; Davies, R. D.; Davis, R. J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Delouis, J.-M.; Démoclès, J.; Désert, F.-X.; Dickinson, C.; Diego, J. M.; Dolag, K.; Dole, H.; Donzelli, S.; Doré, O.; Douspis, M.; Dupac, X.; Efstathiou, G.; Enßlin, T. A.; Eriksen, H. K.; Feroz, F.; Ferragamo, A.; Finelli, F.; Flores-Cacho, I.; Forni, O.; Frailis, M.; Franceschi, E.; Fromenteau, S.; Galeotta, S.; Ganga, K.; Génova-Santos, R. T.; Giard, M.; Giardino, G.; Gilfanov, M.; Giraud-Héraud, Y.; González-Nuevo, J.; Górski, K. M.; Grainge, K. J. B.; Gratton, S.; Gregorio, A.; Groeneboom, N., E.; Gruppuso, A.; Hansen, F. K.; Hanson, D.; Harrison, D.; Hempel, A.; Henrot-Versillé, S.; Hernández-Monteagudo, C.; Herranz, D.; Hildebrandt, S. R.; Hivon, E.; Hobson, M.; Holmes, W. A.; Hornstrup, A.; Hovest, W.; Huffenberger, K. M.; Hurier, G.; Hurley-Walker, N.; Jaffe, A. H.; Jaffe, T. R.; Jones, W. C.; Juvela, M.; Keihänen, E.; Keskitalo, R.; Khamitov, I.; Kisner, T. S.; Kneissl, R.; Knoche, J.; Knox, L.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lähteenmäki, A.; Lamarre, J.-M.; Lasenby, A.; Laureijs, R. J.; Lawrence, C. R.; Leahy, J. P.; Leonardi, R.; León-Tavares, J.; Lesgourgues, J.; Li, C.; Liddle, A.; Liguori, M.; Lilje, P. B.; Linden-Vørnle, M.; López-Caniego, M.; Lubin, P. M.; Macías-Pérez, J. F.; MacTavish, C. J.; Maffei, B.; Maino, D.; Mandolesi, N.; Maris, M.; Marshall, D. J.; Martin, P. G.; Martínez-González, E.; Masi, S.; Massardi, M.; Matarrese, S.; Matthai, F.; Mazzotta, P.; Mei, S.; Meinhold, P. R.; Melchiorri, A.; Melin, J.-B.; Mendes, L.; Mennella, A.; Migliaccio, M.; Mikkelsen, K.; Mitra, S.; Miville-Deschênes, M.-A.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Munshi, D.; Murphy, J. A.; Naselsky, P.; Nastasi, A.; Nati, F.; Natoli, P.; Nesvadba, N. P. H.; Netterfield, C. B.; Nørgaard-Nielsen, H. U.; Noviello, F.; Novikov, D.; Novikov, I.; O'Dwyer, I. J.; Olamaie, M.; Osborne, S.; Oxborrow, C. A.; Paci, F.; Pagano, L.; Pajot, F.; Paoletti, D.; Pasian, F.; Patanchon, G.; Pearson, T. J.; Perdereau, O.; Perotto, L.; Perrott, Y. C.; Perrotta, F.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Ponthieu, N.; Popa, L.; Poutanen, T.; Pratt, G. W.; Prézeau, G.; Prunet, S.; Puget, J.-L.; Rachen, J. P.; Reach, W. T.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Ricciardi, S.; Riller, T.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Roudier, G.; Rowan-Robinson, M.; Rubiño-Martín, J. A.; Rumsey, C.; Rusholme, B.; Sandri, M.; Santos, D.; Saunders, R. D. E.; Savini, G.; Schammel, M. P.; Scott, D.; Seiffert, M. D.; Shellard, E. P. S.; Shimwell, T. W.; Spencer, L. D.; Starck, J.-L.; Stolyarov, V.; Stompor, R.; Streblyanska, A.; Sudiwala, R.; Sunyaev, R.; Sureau, F.; Sutton, D.; Suur-Uski, A.-S.; Sygnet, J.-F.; Tauber, J. A.; Tavagnacco, D.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tramonte, D.; Tristram, M.; Tucci, M.; Tuovinen, J.; Türler, M.; Umana, G.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Vibert, L.; Vielva, P.; Villa, F.; Vittorio, N.; Wade, L. A.; Wandelt, B. D.; White, M.; White, S. D. M.; Yvon, D.; Zacchei, A.; Zonca, A.

    2015-09-01

    We update the all-sky Planck catalogue of 1227 clusters and cluster candidates (PSZ1) published in March 2013, derived from detections of the Sunyaev-Zeldovich (SZ) effect using the first 15.5 months of Planck satellite observations. As an addendum, we deliver an updated version of the PSZ1 catalogue, reporting the further confirmation of 86 Planck-discovered clusters. In total, the PSZ1 now contains 947 confirmed clusters, of which 214 were confirmed as newly discovered clusters through follow-up observations undertaken by the Planck Collaboration. The updated PSZ1 contains redshifts for 913 systems, of which 736 (~ 80.6%) are spectroscopic, and associated mass estimates derived from the Yz mass proxy. We also provide a new SZ quality flag for the remaining 280 candidates. This flag was derived from a novel artificial neural-network classification of the SZ signal. Based on this assessment, the purity of the updated PSZ1 catalogue is estimated to be 94%. In this release, we provide the full updated catalogue and an additional readme file with further information on the Planck SZ detections. The catalogue is only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (ftp://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/581/A14

  6. The British Film Catalogue: 1895-1970.

    Science.gov (United States)

    Gifford, Denis

    This reference book catalogues nearly every commercial film produced in Britain for public entertainment from 1895 to 1970. The entries are listed chronologically by year and month. Each entry is limited to a single film and contains a cross index code number, exhibition date, main title, length, color system, production company, distribution…

  7. CURRENT TRENDS IN CATALOGUING AND THE CHALLENGES ...

    African Journals Online (AJOL)

    resources collected by libraries, results in rich metadata that can be used for many .... in creating timely and high quality records, cataloguers need to develop a ... is a professional function for which there is no substitutes for the human begin.

  8. Compatible Lie Bialgebras

    International Nuclear Information System (INIS)

    Wu Ming-Zhong; Bai Cheng-Ming

    2015-01-01

    A compatible Lie algebra is a pair of Lie algebras such that any linear combination of the two Lie brackets is a Lie bracket. We construct a bialgebra theory of compatible Lie algebras as an analogue of a Lie bialgebra. They can also be regarded as a “compatible version” of Lie bialgebras, that is, a pair of Lie bialgebras such that any linear combination of the two Lie bialgebras is still a Lie bialgebra. Many properties of compatible Lie bialgebras as the “compatible version” of the corresponding properties of Lie bialgebras are presented. In particular, there is a coboundary compatible Lie bialgebra theory with a construction from the classical Yang–Baxter equation in compatible Lie algebras as a combination of two classical Yang–Baxter equations in Lie algebras. Furthermore, a notion of compatible pre-Lie algebra is introduced with an interpretation of its close relation with the classical Yang–Baxter equation in compatible Lie algebras which leads to a construction of the solutions of the latter. As a byproduct, the compatible Lie bialgebras fit into the framework to construct non-constant solutions of the classical Yang–Baxter equation given by Golubchik and Sokolov. (paper)

  9. Generate tri-directional spectra-compatible time histories using HHT method

    Energy Technology Data Exchange (ETDEWEB)

    Li, Bo; Xie, Wei-Chau, E-mail: xie@uwaterloo.ca; Pandey, Mahesh D.

    2016-11-15

    Highlights: • Hilbert–Huang Transform are applied to modify real earthquake records. • Generate tri-directional time histories compatible with target spectra. • Both GRS and FRS are considered as target spectra. • Target spectra with multiple damping ratios are considered. - Abstract: This paper proposes two algorithms to generate spectrum-compatible time histories based on two approaches recommended by USNRC Standard Review Plan 3.7.1. Hilbert–Huang Transform technique is used to analyze frequency contents and amplitudes of seed motions. Through adjusting the frequency contents and amplitudes of seed motions, spectrum-compatible time histories are obtained. The first algorithm is to generate tri-directional time histories compatible with multi-damping target design spectra (ground response spectra or floor response spectra). The second algorithm is to generate tri-directional time histories compatible with single-damping target design spectra. Examples are presented to demonstrate versatility of these two proposed algorithms to generate spectra-compatible time histories.

  10. Generate tri-directional spectra-compatible time histories using HHT method

    International Nuclear Information System (INIS)

    Li, Bo; Xie, Wei-Chau; Pandey, Mahesh D.

    2016-01-01

    Highlights: • Hilbert–Huang Transform are applied to modify real earthquake records. • Generate tri-directional time histories compatible with target spectra. • Both GRS and FRS are considered as target spectra. • Target spectra with multiple damping ratios are considered. - Abstract: This paper proposes two algorithms to generate spectrum-compatible time histories based on two approaches recommended by USNRC Standard Review Plan 3.7.1. Hilbert–Huang Transform technique is used to analyze frequency contents and amplitudes of seed motions. Through adjusting the frequency contents and amplitudes of seed motions, spectrum-compatible time histories are obtained. The first algorithm is to generate tri-directional time histories compatible with multi-damping target design spectra (ground response spectra or floor response spectra). The second algorithm is to generate tri-directional time histories compatible with single-damping target design spectra. Examples are presented to demonstrate versatility of these two proposed algorithms to generate spectra-compatible time histories.

  11. Representation of activity in images using geospatial temporal graphs

    Science.gov (United States)

    Brost, Randolph; McLendon, III, William C.; Parekh, Ojas D.; Rintoul, Mark Daniel; Watson, Jean-Paul; Strip, David R.; Diegert, Carl

    2018-05-01

    Various technologies pertaining to modeling patterns of activity observed in remote sensing images using geospatial-temporal graphs are described herein. Graphs are constructed by representing objects in remote sensing images as nodes, and connecting nodes with undirected edges representing either distance or adjacency relationships between objects and directed edges representing changes in time. Activity patterns may be discerned from the graphs by coding nodes representing persistent objects like buildings differently from nodes representing ephemeral objects like vehicles, and examining the geospatial-temporal relationships of ephemeral nodes within the graph.

  12. The development of a new edition of the gamma-ray spectrum catalogues designed for presentation in electronic format

    Energy Technology Data Exchange (ETDEWEB)

    Heath, R.L. [Idaho National Engineering Lab., Idaho Falls, ID (United States)

    1997-11-01

    New editions of the original Gamma-ray Spectrum Catalogues are being prepared for publication in electronic format. The objective of this program is to produce versions of the Catalogues in CD-ROM format and as an Internet resource. Additions to the original content of the Catalogues will include integrated decay scheme drawings, tables of related decay data, and updated text on the techniques of gamma-ray spectrometry. Related decay data from the Evaluated Nuclear Structure Data File (ENSDF) are then added, and all data converted to the Adobe Acrobat (PDF) format for CD-ROM production and availability on the large-volume Ge detectors, alpha-particle spectra, prompt neutron capture and inelastic scattering gamma-ray spectra, and gross fission product spectra characteristic of fuel cycle waste materials. Characterization of radioactivity in materials is a requirement in many phases of radioactive waste management. Movement, shipping, treatment, all activities which involve handling of mixed waste or TRU categories of waste at all DOE sites will require that measurements and assessment documentation utilize basic nuclear data which are tracable to internationally accepted standard values. This program will involve the identification of data needs unique to the development and application of specialized detector systems for radioactive waste characterization. 8 refs., 8 figs.

  13. The development of a new edition of the gamma-ray spectrum catalogues designed for presentation in electronic format

    International Nuclear Information System (INIS)

    Heath, R.L.

    1997-01-01

    New editions of the original Gamma-ray Spectrum Catalogues are being prepared for publication in electronic format. The objective of this program is to produce versions of the Catalogues in CD-ROM format and as an Internet resource. Additions to the original content of the Catalogues will include integrated decay scheme drawings, tables of related decay data, and updated text on the techniques of gamma-ray spectrometry. Related decay data from the Evaluated Nuclear Structure Data File (ENSDF) are then added, and all data converted to the Adobe Acrobat (PDF) format for CD-ROM production and availability on the large-volume Ge detectors, alpha-particle spectra, prompt neutron capture and inelastic scattering gamma-ray spectra, and gross fission product spectra characteristic of fuel cycle waste materials. Characterization of radioactivity in materials is a requirement in many phases of radioactive waste management. Movement, shipping, treatment, all activities which involve handling of mixed waste or TRU categories of waste at all DOE sites will require that measurements and assessment documentation utilize basic nuclear data which are tracable to internationally accepted standard values. This program will involve the identification of data needs unique to the development and application of specialized detector systems for radioactive waste characterization. 8 refs., 8 figs

  14. Dynamic federations: storage aggregation using open tools and protocols

    CERN Document Server

    Fabrizio Furano, F F; Ricardo Brito da Rocha, R R; Adrien Devresse, A D; Oliver Keeble, O K; Alejandro Alvarez Ayllon, A A

    2012-01-01

    A number of storage elements now offer standard protocol interfaces like NFS 4.1/pNFS and WebDAV, for access to their data repositories, in line with the standardization effort of the European Middleware Initiative (EMI). Also the LCG FileCatalogue (LFC) can offer such features. Here we report on work that seeks to exploit the federation potential of these protocols and build a system that offers a unique view of the storage and metadata ensemble and the possibility of integration of other compatible resources such as those from cloud providers. The challenge, here undertaken by the providers of dCache and DPM, and pragmatically open to other Grid and Cloud storage solutions, is to build such a system while being able to accommodate name translations from existing catalogues (e.g. LFCs), experiment- based metadata catalogues, or stateless algorithmic name translations, also known as ”trivial file catalogues”. Such so-called storage federations of standard protocols-based storage elements give a unique vie...

  15. Nebhydro: Sharing Geospatial Data to Supportwater Management in Nebraska

    Science.gov (United States)

    Kamble, B.; Irmak, A.; Hubbard, K.; Deogun, J.; Dvorak, B.

    2012-12-01

    Recent advances in web-enabled geographical technologies have the potential to make a dramatic impact on development of highly interactive spatial applications on the web for visualization of large-scale geospatial data by water resources and irrigation scientists. Spatial and point scale water resources data visualization are an emerging and challenging application domain. Query based visual explorations of geospatial hydrological data can play an important role in stimulating scientific hypotheses and seeking causal relationships among hydro variables. The Nebraska Hydrological Information System (NebHydro) utilizes ESRI's ArcGIS server technology to increase technological awareness among farmers, irrigation managers and policy makers. Web-based geospatial applications are an effective way to expose scientific hydrological datasets to the research community and the public. NebHydro uses Adobe Flex technology to offer an online visualization and data analysis system for presentation of social and economic data. Internet mapping services is an integrated product of GIS and Internet technologies; it is a favored solution to achieve the interoperability of GIS. The development of Internet based GIS services in the state of Nebraska showcases the benefits of sharing geospatial hydrological data among agencies, resource managers and policy makers. Geospatial hydrological Information (Evapotranspiration from Remote Sensing, vegetation indices (NDVI), USGS Stream gauge data, Climatic data etc.) is generally generated through model simulation (METRIC, SWAP, Linux, Python based scripting etc). Information is compiled into and stored within object oriented relational spatial databases using a geodatabase information model that supports the key data types needed by applications including features, relationships, networks, imagery, terrains, maps and layers. The system provides online access, querying, visualization, and analysis of the hydrological data from several sources

  16. The ASAS-SN bright supernova catalogue - III. 2016

    Science.gov (United States)

    Holoien, T. W.-S.; Brown, J. S.; Stanek, K. Z.; Kochanek, C. S.; Shappee, B. J.; Prieto, J. L.; Dong, Subo; Brimacombe, J.; Bishop, D. W.; Bose, S.; Beacom, J. F.; Bersier, D.; Chen, Ping; Chomiuk, L.; Falco, E.; Godoy-Rivera, D.; Morrell, N.; Pojmanski, G.; Shields, J. V.; Strader, J.; Stritzinger, M. D.; Thompson, Todd A.; Woźniak, P. R.; Bock, G.; Cacella, P.; Conseil, E.; Cruz, I.; Fernandez, J. M.; Kiyota, S.; Koff, R. A.; Krannich, G.; Marples, P.; Masi, G.; Monard, L. A. G.; Nicholls, B.; Nicolas, J.; Post, R. S.; Stone, G.; Wiethoff, W. S.

    2017-11-01

    This catalogue summarizes information for all supernovae discovered by the All-Sky Automated Survey for SuperNovae (ASAS-SN) and all other bright (mpeak ≤ 17), spectroscopically confirmed supernovae discovered in 2016. We then gather the near-infrared through ultraviolet magnitudes of all host galaxies and the offsets of the supernovae from the centres of their hosts from public data bases. We illustrate the results using a sample that now totals 668 supernovae discovered since 2014 May 1, including the supernovae from our previous catalogues, with type distributions closely matching those of the ideal magnitude limited sample from Li et al. This is the third of a series of yearly papers on bright supernovae and their hosts from the ASAS-SN team.

  17. Geospatial Modelling for Micro Zonation of Groundwater Regime in Western Assam, India

    Science.gov (United States)

    Singh, R. P.

    2016-12-01

    Water, most precious natural resource on earth, is vital to sustain the natural system and human civilisation on the earth. The Assam state located in north-eastern part of India has a relatively good source of ground water due to their geographic and physiographic location but there is problem deterioration of groundwater quality causing major health problem in the area. In this study, I tried a integrated study of remote sensing and GIS and chemical analysis of groundwater samples to throw a light over groundwater regime and provides information for decision makers to make sustainable water resource management. The geospatial modelling performed by integrating hydrogeomorphic features. Geomorphology, lineament, Drainage, Landuse/landcover layer were generated through visual interpretation on satellite image (LISS III) based on tone, texture, shape, size, and arrangement of the features. Slope layer was prepared by using SRTM DEM data set .The LULC of the area were categories in to 6 classes of Agricultural field, Forest area ,River, Settlement , Tree-clad area and Wetlands. The geospatial modelling performed through weightage and rank method in GIS, depending on the influence of the features on ground water regime. To Assess the ground water quality of the area 45 groundwater samples have been collected from the field and chemical analysis performed through the standard method in the laboratory. The overall assessment of the ground water quality of the area analyse through Water Quality Index and found that about 70% samples are not potable for drinking purposes due to higher concentration Arsenic, Fluoride and Iron. It appears that, source of all these pollutants geologically and geomorphologically derived. Interpolated layer of Water Quality Index and geospatial modelled Groundwater potential layer provides a holistic view of groundwater scenario and provide direction for better planning and groundwater resource management. Study will be discussed in details

  18. GEO Label Web Services for Dynamic and Effective Communication of Geospatial Metadata Quality

    Science.gov (United States)

    Lush, Victoria; Nüst, Daniel; Bastin, Lucy; Masó, Joan; Lumsden, Jo

    2014-05-01

    We present demonstrations of the GEO label Web services and their integration into a prototype extension of the GEOSS portal (http://scgeoviqua.sapienzaconsulting.com/web/guest/geo_home), the GMU portal (http://gis.csiss.gmu.edu/GADMFS/) and a GeoNetwork catalog application (http://uncertdata.aston.ac.uk:8080/geonetwork/srv/eng/main.home). The GEO label is designed to communicate, and facilitate interrogation of, geospatial quality information with a view to supporting efficient and effective dataset selection on the basis of quality, trustworthiness and fitness for use. The GEO label which we propose was developed and evaluated according to a user-centred design (UCD) approach in order to maximise the likelihood of user acceptance once deployed. The resulting label is dynamically generated from producer metadata in ISO or FDGC format, and incorporates user feedback on dataset usage, ratings and discovered issues, in order to supply a highly informative summary of metadata completeness and quality. The label was easily incorporated into a community portal as part of the GEO Architecture Implementation Programme (AIP-6) and has been successfully integrated into a prototype extension of the GEOSS portal, as well as the popular metadata catalog and editor, GeoNetwork. The design of the GEO label was based on 4 user studies conducted to: (1) elicit initial user requirements; (2) investigate initial user views on the concept of a GEO label and its potential role; (3) evaluate prototype label visualizations; and (4) evaluate and validate physical GEO label prototypes. The results of these studies indicated that users and producers support the concept of a label with drill-down interrogation facility, combining eight geospatial data informational aspects, namely: producer profile, producer comments, lineage information, standards compliance, quality information, user feedback, expert reviews, and citations information. These are delivered as eight facets of a wheel

  19. Stakeholder Alignment and Changing Geospatial Information Capabilities

    Science.gov (United States)

    Winter, S.; Cutcher-Gershenfeld, J.; King, J. L.

    2015-12-01

    Changing geospatial information capabilities can have major economic and social effects on activities such as drought monitoring, weather forecasts, agricultural productivity projections, water and air quality assessments, the effects of forestry practices and so on. Whose interests are served by such changes? Two common mistakes are assuming stability in the community of stakeholders and consistency in stakeholder behavior. Stakeholder communities can reconfigure dramatically as some leave the discussion, others enter, and circumstances shift — all resulting in dynamic points of alignment and misalignment . New stakeholders can bring new interests, and existing stakeholders can change their positions. Stakeholders and their interests need to be be considered as geospatial information capabilities change, but this is easier said than done. New ways of thinking about stakeholder alignment in light of changes in capability are presented.

  20. Provenance metadata gathering and cataloguing of EFIT++ code execution

    Energy Technology Data Exchange (ETDEWEB)

    Lupelli, I., E-mail: ivan.lupelli@ccfe.ac.uk [CCFE, Culham Science Centre, Abingdon, Oxon OX14 3DB (United Kingdom); Muir, D.G.; Appel, L.; Akers, R.; Carr, M. [CCFE, Culham Science Centre, Abingdon, Oxon OX14 3DB (United Kingdom); Abreu, P. [Instituto de Plasmas e Fusão Nuclear, Instituto Superior Técnico, Universidade de Lisboa, 1049-001 Lisboa (Portugal)

    2015-10-15

    Highlights: • An approach for automatic gathering of provenance metadata has been presented. • A provenance metadata catalogue has been created. • The overhead in the code runtime is less than 10%. • The metadata/data size ratio is about ∼20%. • A visualization interface based on Gephi, has been presented. - Abstract: Journal publications, as the final product of research activity, are the result of an extensive complex modeling and data analysis effort. It is of paramount importance, therefore, to capture the origins and derivation of the published data in order to achieve high levels of scientific reproducibility, transparency, internal and external data reuse and dissemination. The consequence of the modern research paradigm is that high performance computing and data management systems, together with metadata cataloguing, have become crucial elements within the nuclear fusion scientific data lifecycle. This paper describes an approach to the task of automatically gathering and cataloguing provenance metadata, currently under development and testing at Culham Center for Fusion Energy. The approach is being applied to a machine-agnostic code that calculates the axisymmetric equilibrium force balance in tokamaks, EFIT++, as a proof of principle test. The proposed approach avoids any code instrumentation or modification. It is based on the observation and monitoring of input preparation, workflow and code execution, system calls, log file data collection and interaction with the version control system. Pre-processing, post-processing, and data export and storage are monitored during the code runtime. Input data signals are captured using a data distribution platform called IDAM. The final objective of the catalogue is to create a complete description of the modeling activity, including user comments, and the relationship between data output, the main experimental database and the execution environment. For an intershot or post-pulse analysis (∼1000

  1. Provenance metadata gathering and cataloguing of EFIT++ code execution

    International Nuclear Information System (INIS)

    Lupelli, I.; Muir, D.G.; Appel, L.; Akers, R.; Carr, M.; Abreu, P.

    2015-01-01

    Highlights: • An approach for automatic gathering of provenance metadata has been presented. • A provenance metadata catalogue has been created. • The overhead in the code runtime is less than 10%. • The metadata/data size ratio is about ∼20%. • A visualization interface based on Gephi, has been presented. - Abstract: Journal publications, as the final product of research activity, are the result of an extensive complex modeling and data analysis effort. It is of paramount importance, therefore, to capture the origins and derivation of the published data in order to achieve high levels of scientific reproducibility, transparency, internal and external data reuse and dissemination. The consequence of the modern research paradigm is that high performance computing and data management systems, together with metadata cataloguing, have become crucial elements within the nuclear fusion scientific data lifecycle. This paper describes an approach to the task of automatically gathering and cataloguing provenance metadata, currently under development and testing at Culham Center for Fusion Energy. The approach is being applied to a machine-agnostic code that calculates the axisymmetric equilibrium force balance in tokamaks, EFIT++, as a proof of principle test. The proposed approach avoids any code instrumentation or modification. It is based on the observation and monitoring of input preparation, workflow and code execution, system calls, log file data collection and interaction with the version control system. Pre-processing, post-processing, and data export and storage are monitored during the code runtime. Input data signals are captured using a data distribution platform called IDAM. The final objective of the catalogue is to create a complete description of the modeling activity, including user comments, and the relationship between data output, the main experimental database and the execution environment. For an intershot or post-pulse analysis (∼1000

  2. VizieR Online Data Catalog: Catalogue of Stellar Spectral Classifications (Skiff, 2009-2016)

    Science.gov (United States)

    Skiff, B. A.

    2014-10-01

    Morgan & Abt 'MKA' paper (1972AJ.....77...35M) not appearing in the two later lists are added. Keenan made continual adjustments to the standards lists up to the time of his death. Thus the late-type standards comprise those marked as high-weight standards in the 1989 Perkins catalogue (1989ApJS...71..245K = III/150); the revised S-type standards in collaboration with Boeshaar (1980ApJS...43..379K); plus the carbon standards and class IIIb 'clump giants' in collaboration with Barnbaum (1996ApJS..105..419B and 1999ApJ...518..859K). In addition, I have made use of the final types by Keenan up to January 2000 shown at the Ohio State Web site (http://www.astronomy.ohio-state.edu/MKCool), accessed in autumn 2003. Though the present file contains all the stars in these lists, only those marked as standards are flagged as such. Garrison's list of MK 'anchor points' might also be consulted in this regard (1994mpyp.conf....3G, and http://www.astro.utoronto.ca/~garrison/mkstds.html). The catalogue includes for the first time results from many large-scale objective-prism spectral surveys done at Case, Stockholm, Crimea, Abastumani, and elsewhere. The stars in these surveys were usually identified only on charts or by other indirect means, and have been overlooked heretofore because of the difficulty in recovering the stars. More complete results from these separate publications, including notes and identifications, have been made available to the CDS, and are kept at the Lowell Observatory ftp area (ftp://ftp.lowell.edu/pub/bas/starcats). Not all of these stars are present in SIMBAD. As a 'living catalogue', an attempt will be made to keep up with current literature, and to extend the indexing of citations back in time. (2 data files).

  3. VizieR Online Data Catalog: Catalogue of Stellar Spectral Classifications (Skiff, 2009-2012)

    Science.gov (United States)

    Skiff, B. A.

    2010-11-01

    Morgan & Abt 'MKA' paper (1972AJ.....77...35M) not appearing in the two later lists are added. Keenan made continual adjustments to the standards lists up to the time of his death. Thus the late-type standards comprise those marked as high-weight standards in the 1989 Perkins catalogue (1989ApJS...71..245K = III/150); the revised S-type standards in collaboration with Boeshaar (1980ApJS...43..379K); plus the carbon standards and class IIIb 'clump giants' in collaboration with Barnbaum (1996ApJS..105..419B and 1999ApJ...518..859K). In addition, I have made use of the final types by Keenan up to January 2000 shown at the Ohio State Web site (http://www.astronomy.ohio-state.edu/MKCool), accessed in autumn 2003. Though the present file contains all the stars in these lists, only those marked as standards are flagged as such. Garrison's list of MK 'anchor points' might also be consulted in this regard (1994mpyp.conf....3G, and http://www.astro.utoronto.ca/~garrison/mkstds.html). The catalogue includes for the first time results from many large-scale objective-prism spectral surveys done at Case, Stockholm, Crimea, Abastumani, and elsewhere. The stars in these surveys were usually identified only on charts or by other indirect means, and have been overlooked heretofore because of the difficulty in recovering the stars. More complete results from these separate publications, including notes and identifications, have been made available to the CDS, and are kept at the Lowell Observatory ftp area (ftp://ftp.lowell.edu/pub/bas/starcats). Not all of these stars are present in SIMBAD. As a 'living catalogue', an attempt will be made to keep up with current literature, and to extend the indexing of citations back in time. (2 data files).

  4. VizieR Online Data Catalog: Catalogue of Stellar Spectral Classifications (Skiff, 2009-2014)

    Science.gov (United States)

    Skiff, B. A.

    2014-10-01

    Morgan & Abt 'MKA' paper (1972AJ.....77...35M) not appearing in the two later lists are added. Keenan made continual adjustments to the standards lists up to the time of his death. Thus the late-type standards comprise those marked as high-weight standards in the 1989 Perkins catalogue (1989ApJS...71..245K = III/150); the revised S-type standards in collaboration with Boeshaar (1980ApJS...43..379K); plus the carbon standards and class IIIb 'clump giants' in collaboration with Barnbaum (1996ApJS..105..419B and 1999ApJ...518..859K). In addition, I have made use of the final types by Keenan up to January 2000 shown at the Ohio State Web site (http://www.astronomy.ohio-state.edu/MKCool), accessed in autumn 2003. Though the present file contains all the stars in these lists, only those marked as standards are flagged as such. Garrison's list of MK 'anchor points' might also be consulted in this regard (1994mpyp.conf....3G, and http://www.astro.utoronto.ca/~garrison/mkstds.html). The catalogue includes for the first time results from many large-scale objective-prism spectral surveys done at Case, Stockholm, Crimea, Abastumani, and elsewhere. The stars in these surveys were usually identified only on charts or by other indirect means, and have been overlooked heretofore because of the difficulty in recovering the stars. More complete results from these separate publications, including notes and identifications, have been made available to the CDS, and are kept at the Lowell Observatory ftp area (ftp://ftp.lowell.edu/pub/bas/starcats). Not all of these stars are present in SIMBAD. As a 'living catalogue', an attempt will be made to keep up with current literature, and to extend the indexing of citations back in time. (2 data files).

  5. VizieR Online Data Catalog: Catalogue of Stellar Spectral Classifications (Skiff, 2009-2013)

    Science.gov (United States)

    Skiff, B. A.

    2013-05-01

    Morgan & Abt 'MKA' paper (1972AJ.....77...35M) not appearing in the two later lists are added. Keenan made continual adjustments to the standards lists up to the time of his death. Thus the late-type standards comprise those marked as high-weight standards in the 1989 Perkins catalogue (1989ApJS...71..245K = III/150); the revised S-type standards in collaboration with Boeshaar (1980ApJS...43..379K); plus the carbon standards and class IIIb 'clump giants' in collaboration with Barnbaum (1996ApJS..105..419B and 1999ApJ...518..859K). In addition, I have made use of the final types by Keenan up to January 2000 shown at the Ohio State Web site (http://www.astronomy.ohio-state.edu/MKCool), accessed in autumn 2003. Though the present file contains all the stars in these lists, only those marked as standards are flagged as such. Garrison's list of MK 'anchor points' might also be consulted in this regard (1994mpyp.conf....3G, and http://www.astro.utoronto.ca/~garrison/mkstds.html). The catalogue includes for the first time results from many large-scale objective-prism spectral surveys done at Case, Stockholm, Crimea, Abastumani, and elsewhere. The stars in these surveys were usually identified only on charts or by other indirect means, and have been overlooked heretofore because of the difficulty in recovering the stars. More complete results from these separate publications, including notes and identifications, have been made available to the CDS, and are kept at the Lowell Observatory ftp area (ftp://ftp.lowell.edu/pub/bas/starcats). Not all of these stars are present in SIMBAD. As a 'living catalogue', an attempt will be made to keep up with current literature, and to extend the indexing of citations back in time. (2 data files).

  6. MyGeoHub: A Collaborative Geospatial Research and Education Platform

    Science.gov (United States)

    Kalyanam, R.; Zhao, L.; Biehl, L. L.; Song, C. X.; Merwade, V.; Villoria, N.

    2017-12-01

    Scientific research is increasingly collaborative and globally distributed; research groups now rely on web-based scientific tools and data management systems to simplify their day-to-day collaborative workflows. However, such tools often lack seamless interfaces, requiring researchers to contend with manual data transfers, annotation and sharing. MyGeoHub is a web platform that supports out-of-the-box, seamless workflows involving data ingestion, metadata extraction, analysis, sharing and publication. MyGeoHub is built on the HUBzero cyberinfrastructure platform and adds general-purpose software building blocks (GABBs), for geospatial data management, visualization and analysis. A data management building block iData, processes geospatial files, extracting metadata for keyword and map-based search while enabling quick previews. iData is pervasive, allowing access through a web interface, scientific tools on MyGeoHub or even mobile field devices via a data service API. GABBs includes a Python map library as well as map widgets that in a few lines of code, generate complete geospatial visualization web interfaces for scientific tools. GABBs also includes powerful tools that can be used with no programming effort. The GeoBuilder tool provides an intuitive wizard for importing multi-variable, geo-located time series data (typical of sensor readings, GPS trackers) to build visualizations supporting data filtering and plotting. MyGeoHub has been used in tutorials at scientific conferences and educational activities for K-12 students. MyGeoHub is also constantly evolving; the recent addition of Jupyter and R Shiny notebook environments enable reproducible, richly interactive geospatial analyses and applications ranging from simple pre-processing to published tools. MyGeoHub is not a monolithic geospatial science gateway, instead it supports diverse needs ranging from just a feature-rich data management system, to complex scientific tools and workflows.

  7. DIGI-vis: Distributed interactive geospatial information visualization

    KAUST Repository

    Ponto, Kevin; Kuester, Falk

    2010-01-01

    data sets. We propose a distributed data gathering and visualization system that allows researchers to view these data at hundreds of megapixels simultaneously. This system allows scientists to view real-time geospatial information at unprecedented

  8. Informal information for web-based engineering catalogues

    Science.gov (United States)

    Allen, Richard D.; Culley, Stephen J.; Hicks, Ben J.

    2001-10-01

    Success is highly dependent on the ability of a company to efficiently produce optimal designs. In order to achieve this companies must minimize time to market and possess the ability to make fully informed decisions at the early phase of the design process. Such decisions may include the choice of component and suppliers, as well as cost and maintenance considerations. Computer modeling and electronic catalogues are becoming the preferred medium for the selection and design of mechanical components. In utilizing these techniques, the designer demands the capability to identify, evaluate and select mechanical components both quantitatively and qualitatively. Quantitative decisions generally encompass performance data included in the formal catalogue representation. It is in the area of qualitative decisions that the use of what the authors call 'Informal Information' is of crucial importance. Thus, 'Informal Information' must often be incorporated into the selection process and selection systems. This would enable more informed decisions to be made quicker, without the need for information retrieval via discussion with colleagues in the design environment. This paper provides an overview of the use of electronic information in the design of mechanical systems, including a discussion of limitations of current technology. The importance of Informal Information is discussed and the requirements for association with web based electronic catalogues are developed. This system is based on a flexible XML schema and enables the storage, classification and recall of Informal Information packets. Furthermore, a strategy for the inclusion of Informal Information is proposed, and an example case is used to illustrate the benefits.

  9. 3D geospatial visualizations: Animation and motion effects on spatial objects

    Science.gov (United States)

    Evangelidis, Konstantinos; Papadopoulos, Theofilos; Papatheodorou, Konstantinos; Mastorokostas, Paris; Hilas, Constantinos

    2018-02-01

    Digital Elevation Models (DEMs), in combination with high quality raster graphics provide realistic three-dimensional (3D) representations of the globe (virtual globe) and amazing navigation experience over the terrain through earth browsers. In addition, the adoption of interoperable geospatial mark-up languages (e.g. KML) and open programming libraries (Javascript) makes it also possible to create 3D spatial objects and convey on them the sensation of any type of texture by utilizing open 3D representation models (e.g. Collada). One step beyond, by employing WebGL frameworks (e.g. Cesium.js, three.js) animation and motion effects are attributed on 3D models. However, major GIS-based functionalities in combination with all the above mentioned visualization capabilities such as for example animation effects on selected areas of the terrain texture (e.g. sea waves) as well as motion effects on 3D objects moving in dynamically defined georeferenced terrain paths (e.g. the motion of an animal over a hill, or of a big fish in an ocean etc.) are not widely supported at least by open geospatial applications or development frameworks. Towards this we developed and made available to the research community, an open geospatial software application prototype that provides high level capabilities for dynamically creating user defined virtual geospatial worlds populated by selected animated and moving 3D models on user specified locations, paths and areas. At the same time, the generated code may enhance existing open visualization frameworks and programming libraries dealing with 3D simulations, with the geospatial aspect of a virtual world.

  10. NASA space geodesy program: Catalogue of site information

    Science.gov (United States)

    Bryant, M. A.; Noll, C. E.

    1993-01-01

    This is the first edition of the NASA Space Geodesy Program: Catalogue of Site Information. This catalogue supersedes all previous versions of the Crustal Dynamics Project: Catalogue of Site Information, last published in May 1989. This document is prepared under the direction of the Space Geodesy and Altimetry Projects Office (SGAPO), Code 920.1, Goddard Space Flight Center. SGAPO has assumed the responsibilities of the Crustal Dynamics Project, which officially ended December 31, 1991. The catalog contains information on all NASA supported sites as well as sites from cooperating international partners. This catalog is designed to provde descriptions and occupation histories of high-accuracy geodetic measuring sites employing space-related techniques. The emphasis of the catalog has been in the past, and continues to be with this edition, station information for facilities and remote locations utilizing the Satellite Laser Ranging (SLR), Lunar Laser Ranging (LLR), and Very Long Baseline Interferometry (VLBI) techniques. With the proliferation of high-quality Global Positioning System (GPS) receivers and Doppler Orbitography and Radiopositioning Integrated by Satellite (DORIS) transponders, many co-located at established SLR and VLBI observatories, the requirement for accurate station and localized survey information for an ever broadening base of scientists and engineers has been recognized. It is our objective to provide accurate station information to scientific groups interested in these facilities.

  11. Process and results of the development of an ICNP® Catalogue for Cancer Pain

    Directory of Open Access Journals (Sweden)

    Marisaulina Wanderley Abrantes de Carvalho

    2013-10-01

    Full Text Available This was a methodological study conducted to describe the process and results of the development of an International Classification for Nursing Practice (ICNP® Catalogue for Cancer Pain. According to the International Council of Nurses (ICN, this catalogue contains a subset of nursing diagnoses, outcomes, and interventions to document the implementation of the nursing process in cancer patients. This catalogue was developed in several steps according to the guidelines recommended by the ICN. As a result, 68 statements on nursing diagnoses/outcomes were obtained, which were classified according to the theoretical model for nursing care related to cancer pain into physical (28, psychological (29, and sociocultural and spiritual (11 aspects. A total of 116 corresponding nursing interventions were obtained. The proposed ICNP® Catalogue for Cancer Pain aims to provide safe and systematic orientation to nurses who work in this field, thus improving the quality of patient care and facilitating the performance of the nursing process.

  12. Vienna International Centre Library Film and Video Catalogue: Peaceful applications of nuclear energy 1928-1998

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1999-12-31

    The catalogue lists films and videos which are available on free loan from Vienna International Centre Library for educational, non-commercial, non-profit showings involving no admission charges or appeals for funds. Much of the material listed has been donated to the IAEA by the Governments of Member States. The items are arranged in the catalogue by number. The catalogue also includes a title index and a subject index

  13. Vienna International Centre Library Film and Video Catalogue: Peaceful applications of nuclear energy 1928-1998

    International Nuclear Information System (INIS)

    1998-01-01

    The catalogue lists films and videos which are available on free loan from Vienna International Centre Library for educational, non-commercial, non-profit showings involving no admission charges or appeals for funds. Much of the material listed has been donated to the IAEA by the Governments of Member States. The items are arranged in the catalogue by number. The catalogue also includes a title index and a subject index

  14. NativeView: A Geospatial Curriculum for Native Nation Building

    Science.gov (United States)

    Rattling Leaf, J.

    2007-12-01

    In the spirit of collaboration and reciprocity, James Rattling Leaf of Sinte Gleska University on the Rosebud Reservation of South Dakota will present recent developments, experiences, insights and a vision for education in Indian Country. As a thirty-year young institution, Sinte Gleska University is founded by a strong vision of ancestral leadership and the values of the Lakota Way of Life. Sinte Gleska University (SGU) has initiated the development of a Geospatial Education Curriculum project. NativeView: A Geospatial Curriculum for Native Nation Building is a two-year project that entails a disciplined approach towards the development of a relevant Geospatial academic curriculum. This project is designed to meet the educational and land management needs of the Rosebud Lakota Tribe through the utilization of Geographic Information Systems (GIS), Remote Sensing (RS) and Global Positioning Systems (GPS). In conjunction with the strategy and progress of this academic project, a formal presentation and demonstration of the SGU based Geospatial software RezMapper software will exemplify an innovative example of state of the art information technology. RezMapper is an interactive CD software package focused toward the 21 Lakota communities on the Rosebud Reservation that utilizes an ingenious concept of multimedia mapping and state of the art data compression and presentation. This ongoing development utilizes geographic data, imagery from space, historical aerial photography and cultural features such as historic Lakota documents, language, song, video and historical photographs in a multimedia fashion. As a tangible product, RezMapper will be a project deliverable tool for use in the classroom and to a broad range of learners.

  15. The new geospatial tools: global transparency enhancing safeguards verification

    International Nuclear Information System (INIS)

    Pabian, Frank Vincent

    2010-01-01

    This paper focuses on the importance and potential role of the new, freely available, geospatial tools for enhancing IAEA safeguards and how, together with commercial satellite imagery, they can be used to promote 'all-source synergy'. As additional 'open sources', these new geospatial tools have heralded a new era of 'global transparency' and they can be used to substantially augment existing information-driven safeguards gathering techniques, procedures, and analyses in the remote detection of undeclared facilities, as well as support ongoing monitoring and verification of various treaty (e.g., NPT, FMCT) relevant activities and programs. As an illustration of how these new geospatial tools may be applied, an original exemplar case study provides how it is possible to derive value-added follow-up information on some recent public media reporting of a former clandestine underground plutonium production complex (now being converted to a 'Tourist Attraction' given the site's abandonment by China in the early 1980s). That open source media reporting, when combined with subsequent commentary found in various Internet-based Blogs and Wikis, led to independent verification of the reporting with additional ground truth via 'crowdsourcing' (tourist photos as found on 'social networking' venues like Google Earth's Panoramio layer and Twitter). Confirmation of the precise geospatial location of the site (along with a more complete facility characterization incorporating 3-D Modeling and visualization) was only made possible following the acquisition of higher resolution commercial satellite imagery that could be correlated with the reporting, ground photos, and an interior diagram, through original imagery analysis of the overhead imagery.

  16. The sixth catalogue of galactic Wolf-Rayet stars, their past and present

    International Nuclear Information System (INIS)

    Hucht, K.A. van der; Conti, P.S.; Lundstroem, I.; Stenholm, B.

    1981-01-01

    This paper presents the Sixth Catalogue of galactic Wolf-Rayet stars (Pop I), a short history on the five earlier WR catalogues, improved spectral classification, finding charts, a discussion on related objects, and a review of the current statur of Wolf-Rayet star research. The appendix presents a bibliography on most of the Wolf-Rayet literature published since 1867. (orig.)

  17. Catalogue of high-mass X-ray binaries in the Galaxy (4th edition)

    NARCIS (Netherlands)

    Liu, Q.Z.; van Paradijs, J.; van den Heuvel, E.P.J.

    2006-01-01

    We present a new edition of the catalogue of high-mass X-ray binaries in the Galaxy. The catalogue contains source name(s), coordinates, finding chart, X-ray luminosity, system parameters, and stellar parameters of the components and other characteristic properties of 114 high-mass X-ray binaries,

  18. A Practice Approach of Multi-source Geospatial Data Integration for Web-based Geoinformation Services

    Science.gov (United States)

    Huang, W.; Jiang, J.; Zha, Z.; Zhang, H.; Wang, C.; Zhang, J.

    2014-04-01

    Geospatial data resources are the foundation of the construction of geo portal which is designed to provide online geoinformation services for the government, enterprise and public. It is vital to keep geospatial data fresh, accurate and comprehensive in order to satisfy the requirements of application and development of geographic location, route navigation, geo search and so on. One of the major problems we are facing is data acquisition. For us, integrating multi-sources geospatial data is the mainly means of data acquisition. This paper introduced a practice integration approach of multi-source geospatial data with different data model, structure and format, which provided the construction of National Geospatial Information Service Platform of China (NGISP) with effective technical supports. NGISP is the China's official geo portal which provides online geoinformation services based on internet, e-government network and classified network. Within the NGISP architecture, there are three kinds of nodes: national, provincial and municipal. Therefore, the geospatial data is from these nodes and the different datasets are heterogeneous. According to the results of analysis of the heterogeneous datasets, the first thing we do is to define the basic principles of data fusion, including following aspects: 1. location precision; 2.geometric representation; 3. up-to-date state; 4. attribute values; and 5. spatial relationship. Then the technical procedure is researched and the method that used to process different categories of features such as road, railway, boundary, river, settlement and building is proposed based on the principles. A case study in Jiangsu province demonstrated the applicability of the principle, procedure and method of multi-source geospatial data integration.

  19. Recent innovation of geospatial information technology to support disaster risk management and responses

    Science.gov (United States)

    Une, Hiroshi; Nakano, Takayuki

    2018-05-01

    Geographic location is one of the most fundamental and indispensable information elements in the field of disaster response and prevention. For example, in the case of the Tohoku Earthquake in 2011, aerial photos taken immediately after the earthquake greatly improved information sharing among different government offices and facilitated rescue and recovery operations, and maps prepared after the disaster assisted in the rapid reconstruction of affected local communities. Thanks to the recent development of geospatial information technology, this information has become more essential for disaster response activities. Advancements in web mapping technology allows us to better understand the situation by overlaying various location-specific data on base maps on the web and specifying the areas on which activities should be focused. Through 3-D modelling technology, we can have a more realistic understanding of the relationship between disaster and topography. Geospatial information technology can sup-port proper preparation and emergency responses against disasters by individuals and local communities through hazard mapping and other information services using mobile devices. Thus, geospatial information technology is playing a more vital role on all stages of disaster risk management and responses. In acknowledging geospatial information's vital role in disaster risk reduction, the Sendai Framework for Disaster Risk Reduction 2015-2030, adopted at the Third United Nations World Conference on Disaster Risk Reduction, repeatedly reveals the importance of utilizing geospatial information technology for disaster risk reduction. This presentation aims to report the recent practical applications of geospatial information technology for disaster risk management and responses.

  20. Developing a distributed HTML5-based search engine for geospatial resource discovery

    Science.gov (United States)

    ZHOU, N.; XIA, J.; Nebert, D.; Yang, C.; Gui, Z.; Liu, K.

    2013-12-01

    With explosive growth of data, Geospatial Cyberinfrastructure(GCI) components are developed to manage geospatial resources, such as data discovery and data publishing. However, the efficiency of geospatial resources discovery is still challenging in that: (1) existing GCIs are usually developed for users of specific domains. Users may have to visit a number of GCIs to find appropriate resources; (2) The complexity of decentralized network environment usually results in slow response and pool user experience; (3) Users who use different browsers and devices may have very different user experiences because of the diversity of front-end platforms (e.g. Silverlight, Flash or HTML). To address these issues, we developed a distributed and HTML5-based search engine. Specifically, (1)the search engine adopts a brokering approach to retrieve geospatial metadata from various and distributed GCIs; (2) the asynchronous record retrieval mode enhances the search performance and user interactivity; (3) the search engine based on HTML5 is able to provide unified access capabilities for users with different devices (e.g. tablet and smartphone).

  1. Geospatial distribution modeling and determining suitability of groundwater quality for irrigation purpose using geospatial methods and water quality index (WQI) in Northern Ethiopia

    Science.gov (United States)

    Gidey, Amanuel

    2018-06-01

    Determining suitability and vulnerability of groundwater quality for irrigation use is a key alarm and first aid for careful management of groundwater resources to diminish the impacts on irrigation. This study was conducted to determine the overall suitability of groundwater quality for irrigation use and to generate their spatial distribution maps in Elala catchment, Northern Ethiopia. Thirty-nine groundwater samples were collected to analyze and map the water quality variables. Atomic absorption spectrophotometer, ultraviolet spectrophotometer, titration and calculation methods were used for laboratory groundwater quality analysis. Arc GIS, geospatial analysis tools, semivariogram model types and interpolation methods were used to generate geospatial distribution maps. Twelve and eight water quality variables were used to produce weighted overlay and irrigation water quality index models, respectively. Root-mean-square error, mean square error, absolute square error, mean error, root-mean-square standardized error, measured values versus predicted values were used for cross-validation. The overall weighted overlay model result showed that 146 km2 areas are highly suitable, 135 km2 moderately suitable and 60 km2 area unsuitable for irrigation use. The result of irrigation water quality index confirms 10.26% with no restriction, 23.08% with low restriction, 20.51% with moderate restriction, 15.38% with high restriction and 30.76% with the severe restriction for irrigation use. GIS and irrigation water quality index are better methods for irrigation water resources management to achieve a full yield irrigation production to improve food security and to sustain it for a long period, to avoid the possibility of increasing environmental problems for the future generation.

  2. The Efficacy of Educative Curriculum Materials to Support Geospatial Science Pedagogical Content Knowledge

    Science.gov (United States)

    Bodzin, Alec; Peffer, Tamara; Kulo, Violet

    2012-01-01

    Teaching and learning about geospatial aspects of energy resource issues requires that science teachers apply effective science pedagogical approaches to implement geospatial technologies into classroom instruction. To address this need, we designed educative curriculum materials as an integral part of a comprehensive middle school energy…

  3. Geospatial Information Categories Mapping in a Cross-lingual Environment: A Case Study of “Surface Water” Categories in Chinese and American Topographic Maps

    Directory of Open Access Journals (Sweden)

    Xi Kuai

    2016-06-01

    Full Text Available The need for integrating geospatial information (GI data from various heterogeneous sources has seen increased importance for geographic information system (GIS interoperability. Using domain ontologies to clarify and integrate the semantics of data is considered as a crucial step for successful semantic integration in the GI domain. Nevertheless, mechanisms are still needed to facilitate semantic mapping between GI ontologies described in different natural languages. This research establishes a formal ontology model for cross-lingual geospatial information ontology mapping. By first extracting semantic primitives from a free-text definition of categories in two GI classification standards with different natural languages, an ontology-driven approach is used, and a formal ontology model is established to formally represent these semantic primitives into semantic statements, in which the spatial-related properties and relations are considered as crucial statements for the representation and identification of the semantics of the GI categories. Then, an algorithm is proposed to compare these semantic statements in a cross-lingual environment. We further design a similarity calculation algorithm based on the proposed formal ontology model to distance the semantic similarities and identify the mapping relationships between categories. In particular, we work with two GI classification standards for Chinese and American topographic maps. The experimental results demonstrate the feasibility and reliability of the proposed model for cross-lingual geospatial information ontology mapping.

  4. lawn: An R client for the Turf JavaScript Library for Geospatial Analysis

    Science.gov (United States)

    lawn is an R package to provide access to the geospatial analysis capabilities in the Turf javascript library. Turf expects data in GeoJSON format. Given that many datasets are now available natively in GeoJSON providing an easier method for conducting geospatial analyses on thes...

  5. Promenade Among Words and Things: The Gallery as Catalogue, the Catalogue as Gallery

    Directory of Open Access Journals (Sweden)

    Mari Lending

    2015-12-01

    Full Text Available In the mid nineteenth century new casting techniques allowed for the production of huge architectural fragments. Well-selected collections could ideally display perfect series in galleries in which the visitor could wander among monuments and experience architecture history on full scale. The disembodied material of plaster was considered capable of embodying a number of modern historical taxonomies and aesthetical programs, most importantly chronology, comparison, style, and evolution. Veritable showcases of historicism, the casts could illustrate in spatial arrangements new conceptions on the history, contemporaneity and future of architecture. Plaster cast became a main medium in which to publish antiquities as novelties for grand audiences, taking the printed and published beyond the two-dimensional space of words and images. However, due to the increasing market of casts and their sheere size and weight, the reproductions as mounted in the galleries often behaved as unruly as architecture does outside curatorial control. In the end only the catalogues, the paper versions of these imaginary museums were capable to create the orders that their plaster referents constantly aspired to destroy. An important chapter in the history of the architecture museum these plaster monuments belong to a part of architectural print culture in which catalogues were curated and galleries edited. Metaphors drawn from the realm of writing saturated the discourse on the display of casts. Images and texts fluctuated and the image-objects were compared to books, paper, pages, documents and libraries but above all to illustrations inviting promenades in time and space.

  6. Developing Daily Quantitative Damage Estimates From Geospatial Layers To Support Post Event Recovery

    Science.gov (United States)

    Woods, B. K.; Wei, L. H.; Connor, T. C.

    2014-12-01

    With the growth of natural hazard data available in near real-time it is increasingly feasible to deliver damage estimates caused by natural disasters. These estimates can be used in disaster management setting or by commercial entities to optimize the deployment of resources and/or routing of goods and materials. This work outlines an end-to-end, modular process to generate estimates of damage caused by severe weather. The processing stream consists of five generic components: 1) Hazard modules that provide quantitate data layers for each peril. 2) Standardized methods to map the hazard data to an exposure layer based on atomic geospatial blocks. 3) Peril-specific damage functions that compute damage metrics at the atomic geospatial block level. 4) Standardized data aggregators, which map damage to user-specific geometries. 5) Data dissemination modules, which provide resulting damage estimates in a variety of output forms. This presentation provides a description of this generic tool set, and an illustrated example using HWRF-based hazard data for Hurricane Arthur (2014). In this example, the Python-based real-time processing ingests GRIB2 output from the HWRF numerical model, dynamically downscales it in conjunctions with a land cover database using a multiprocessing pool, and a just-in-time compiler (JIT). The resulting wind fields are contoured, and ingested into a PostGIS database using OGR. Finally, the damage estimates are calculated at the atomic block level and aggregated to user-defined regions using PostgreSQL queries to construct application specific tabular and graphics output.

  7. The WATCH solar X-ray burst catalogue

    DEFF Research Database (Denmark)

    Crosby, N.; Lund, Niels; Vilmer, N.

    1998-01-01

    The WATCH experiment aboard the GRANAT satellite provides observations of the Sun in the deka-keV range covering the years 1990 through mid-1992. An introduction to the experiment is given followed by an explanation of how the WATCH solar burst catalogue was created. The different parameters list...

  8. A Geospatial Data Recommender System based on Metadata and User Behaviour

    Science.gov (United States)

    Li, Y.; Jiang, Y.; Yang, C. P.; Armstrong, E. M.; Huang, T.; Moroni, D. F.; Finch, C. J.; McGibbney, L. J.

    2017-12-01

    Earth observations are produced in a fast velocity through real time sensors, reaching tera- to peta- bytes of geospatial data daily. Discovering and accessing the right data from the massive geospatial data is like finding needle in the haystack. To help researchers find the right data for study and decision support, quite a lot of research focusing on improving search performance have been proposed including recommendation algorithm. However, few papers have discussed the way to implement a recommendation algorithm in geospatial data retrieval system. In order to address this problem, we propose a recommendation engine to improve discovering relevant geospatial data by mining and utilizing metadata and user behavior data: 1) metadata based recommendation considers the correlation of each attribute (i.e., spatiotemporal, categorical, and ordinal) to data to be found. In particular, phrase extraction method is used to improve the accuracy of the description similarity; 2) user behavior data are utilized to predict the interest of a user through collaborative filtering; 3) an integration method is designed to combine the results of the above two methods to achieve better recommendation Experiments show that in the hybrid recommendation list, the all the precisions are larger than 0.8 from position 1 to 10.

  9. Dark Energy Survey Year 1 Results: galaxy mock catalogues for BAO

    Energy Technology Data Exchange (ETDEWEB)

    Avila, S.; et al.

    2017-12-17

    Mock catalogues are a crucial tool in the analysis of galaxy surveys data, both for the accurate computation of covariance matrices, and for the optimisation of analysis methodology and validation of data sets. In this paper, we present a set of 1800 galaxy mock catalogues designed to match the Dark Energy Survey Year-1 BAO sample (Crocce et al. 2017) in abundance, observational volume, redshift distribution and uncertainty, and redshift dependent clustering. The simulated samples were built upon HALOGEN (Avila et al. 2015) halo catalogues, based on a $2LPT$ density field with an exponential bias. For each of them, a lightcone is constructed by the superposition of snapshots in the redshift range $0.45catalogues and compare their clustering to the data using the angular correlation function $ w(\\theta)$, the comoving transverse separation clustering $\\xi_{\\mu<0.8}(s_{\\perp})$ and the angular power spectrum $C_\\ell$.

  10. VISA: AN AUTOMATIC AWARE AND VISUAL AIDS MECHANISM FOR IMPROVING THE CORRECT USE OF GEOSPATIAL DATA

    Directory of Open Access Journals (Sweden)

    J. H. Hong

    2016-06-01

    Full Text Available With the fast growth of internet-based sharing mechanism and OpenGIS technology, users nowadays enjoy the luxury to quickly locate and access a variety of geospatial data for the tasks at hands. While this sharing innovation tremendously expand the possibility of application and reduce the development cost, users nevertheless have to deal with all kinds of “differences” implicitly hidden behind the acquired georesources. We argue the next generation of GIS-based environment, regardless internet-based or not, must have built-in knowledge to automatically and correctly assess the fitness of data use and present the analyzed results to users in an intuitive and meaningful way. The VISA approach proposed in this paper refer to four different types of visual aids that can be respectively used for addressing analyzed results, namely, virtual layer, informative window, symbol transformation and augmented TOC. The VISA-enabled interface works in an automatic-aware fashion, where the standardized metadata serve as the known facts about the selected geospatial resources, algorithms for analyzing the differences of temporality and quality of the geospatial resources were designed and the transformation of analyzed results into visual aids were automatically executed. It successfully presents a new way for bridging the communication gaps between systems and users. GIS has been long seen as a powerful integration tool, but its achievements would be highly restricted if it fails to provide a friendly and correct working platform.

  11. A Geospatial Decision Support System Toolkit, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — We propose to build and commercialize a working prototype Geospatial Decision Support Toolkit (GeoKit). GeoKit will enable scientists, agencies, and stakeholders to...

  12. Nuclear Knowledge Management Case Studies Catalogue “NKM CSC”

    International Nuclear Information System (INIS)

    Atieh, T.

    2016-01-01

    Full text: Over the past several years, many nuclear organizations in IAEA’s Member States have accumulated considerable experiences and achievements in the development and application of nuclear knowledge management (NKM) methodology and tools to improve their organizational performance. The IAEA NKM Section has initiated a project entitled “NKM Case Studies Catalogue (NKM CSC)” to capture and document, as well as preserve NKM experience and facilitate its sharing among NKM practitioners and experts. This is done through collection and preservation of information of relevant experiential knowledge in “case study” format. The catalogue will therefore support community of practice mechanisms. An input template is currently under development and will be used to help contributors in Member States who are providing concise set of information about their respective case studies. This information will be made searchable and easily retrievable through a platform that supports collaboration among NKM practitioners and experts. It is planned to launch the Nuclear Knowledge Management Case Studies Catalogue “NKM CSC” at the occasion of the “Third International Conference on Nuclear Knowledge Management—Challenges and Approaches, 7-–11 November 2016, Vienna, Austria”, and to include the accepted case studies submitted to this Conference. (author

  13. Resolving taxonmic discrepancies: Role of Electronic Catalogues of Known Organisms

    Directory of Open Access Journals (Sweden)

    Vishwas Chavan

    2005-01-01

    Full Text Available There is a disparity in availability of nomenclature change literature to the taxonomists of the developing world and availability of taxonomic papers published by developing world scientists to their counterparts in developed part of the globe. This has resulted in several discrepancies in the naming of organisms. Development of electronic catalogues of names of known organisms would help in pointing out these issues. We have attempted to highlight a few of such discrepancies found while developing IndFauna, an electronic catalogue of known Indian fauna and comparing it with existing global and regional databases.Full Text: PDF

  14. The new geospatial tools: global transparency enhancing safeguards verification

    Energy Technology Data Exchange (ETDEWEB)

    Pabian, Frank Vincent [Los Alamos National Laboratory

    2010-09-16

    This paper focuses on the importance and potential role of the new, freely available, geospatial tools for enhancing IAEA safeguards and how, together with commercial satellite imagery, they can be used to promote 'all-source synergy'. As additional 'open sources', these new geospatial tools have heralded a new era of 'global transparency' and they can be used to substantially augment existing information-driven safeguards gathering techniques, procedures, and analyses in the remote detection of undeclared facilities, as well as support ongoing monitoring and verification of various treaty (e.g., NPT, FMCT) relevant activities and programs. As an illustration of how these new geospatial tools may be applied, an original exemplar case study provides how it is possible to derive value-added follow-up information on some recent public media reporting of a former clandestine underground plutonium production complex (now being converted to a 'Tourist Attraction' given the site's abandonment by China in the early 1980s). That open source media reporting, when combined with subsequent commentary found in various Internet-based Blogs and Wikis, led to independent verification of the reporting with additional ground truth via 'crowdsourcing' (tourist photos as found on 'social networking' venues like Google Earth's Panoramio layer and Twitter). Confirmation of the precise geospatial location of the site (along with a more complete facility characterization incorporating 3-D Modeling and visualization) was only made possible following the acquisition of higher resolution commercial satellite imagery that could be correlated with the reporting, ground photos, and an interior diagram, through original imagery analysis of the overhead imagery.

  15. Catalogue Creation for Space Situational Awareness with Optical Sensors

    Science.gov (United States)

    Hobson, T.; Clarkson, I.; Bessell, T.; Rutten, M.; Gordon, N.; Moretti, N.; Morreale, B.

    2016-09-01

    In order to safeguard the continued use of space-based technologies, effective monitoring and tracking of man-made resident space objects (RSOs) is paramount. The diverse characteristics, behaviours and trajectories of RSOs make space surveillance a challenging application of the discipline that is tracking and surveillance. When surveillance systems are faced with non-canonical scenarios, it is common for human operators to intervene while researchers adapt and extend traditional tracking techniques in search of a solution. A complementary strategy for improving the robustness of space surveillance systems is to place greater emphasis on the anticipation of uncertainty. Namely, give the system the intelligence necessary to autonomously react to unforeseen events and to intelligently and appropriately act on tenuous information rather than discard it. In this paper we build from our 2015 campaign and describe the progression of a low-cost intelligent space surveillance system capable of autonomously cataloguing and maintaining track of RSOs. It currently exploits robotic electro-optical sensors, high-fidelity state-estimation and propagation as well as constrained initial orbit determination (IOD) to intelligently and adaptively manage its sensors in order to maintain an accurate catalogue of RSOs. In a step towards fully autonomous cataloguing, the system has been tasked with maintaining surveillance of a portion of the geosynchronous (GEO) belt. Using a combination of survey and track-refinement modes, the system is capable of maintaining a track of known RSOs and initiating tracks on previously unknown objects. Uniquely, due to the use of high-fidelity representations of a target's state uncertainty, as few as two images of previously unknown RSOs may be used to subsequently initiate autonomous search and reacquisition. To achieve this capability, particularly within the congested environment of the GEO-belt, we use a constrained admissible region (CAR) to

  16. Prototyping an online wetland ecosystem services model using open model sharing standards

    Science.gov (United States)

    Feng, M.; Liu, S.; Euliss, N.H.; Young, Caitlin; Mushet, D.M.

    2011-01-01

    Great interest currently exists for developing ecosystem models to forecast how ecosystem services may change under alternative land use and climate futures. Ecosystem services are diverse and include supporting services or functions (e.g., primary production, nutrient cycling), provisioning services (e.g., wildlife, groundwater), regulating services (e.g., water purification, floodwater retention), and even cultural services (e.g., ecotourism, cultural heritage). Hence, the knowledge base necessary to quantify ecosystem services is broad and derived from many diverse scientific disciplines. Building the required interdisciplinary models is especially challenging as modelers from different locations and times may develop the disciplinary models needed for ecosystem simulations, and these models must be identified and made accessible to the interdisciplinary simulation. Additional difficulties include inconsistent data structures, formats, and metadata required by geospatial models as well as limitations on computing, storage, and connectivity. Traditional standalone and closed network systems cannot fully support sharing and integrating interdisciplinary geospatial models from variant sources. To address this need, we developed an approach to openly share and access geospatial computational models using distributed Geographic Information System (GIS) techniques and open geospatial standards. We included a means to share computational models compliant with Open Geospatial Consortium (OGC) Web Processing Services (WPS) standard to ensure modelers have an efficient and simplified means to publish new models. To demonstrate our approach, we developed five disciplinary models that can be integrated and shared to simulate a few of the ecosystem services (e.g., water storage, waterfowl breeding) that are provided by wetlands in the Prairie Pothole Region (PPR) of North America.

  17. Incentive Compatibility

    OpenAIRE

    Ledyard, John O.

    1987-01-01

    Incentive compatibility is described and discussed. A summary of the current state of understanding is provided. Key words are: incentive compatibility, game theory, implementation, mechanism, Bayes, Nash, and revelation.

  18. Technologies Connotation and Developing Characteristics of Open Geospatial Information Platform

    Directory of Open Access Journals (Sweden)

    GUO Renzhong

    2016-02-01

    Full Text Available Based on the background of developments of surveying,mapping and geoinformation,aimed at the demands of data fusion,real-time sharing,in-depth processing and personalization,this paper analyzes significant features of geo-spatial service in digital city,focuses on theory,method and key techniques of open environment of cloud computing,multi-path data updating,full-scale urban geocoding,multi-source spatial data integration,adaptive geo-processing and adaptive Web mapping.As the basis for it,the Open Geospatial information platform is developed,and successfully implicated in digital Shenzhen.

  19. Assessing the socioeconomic impact and value of open geospatial information

    Science.gov (United States)

    Pearlman, Francoise; Pearlman, Jay; Bernknopf, Richard; Coote, Andrew; Craglia, Massimo; Friedl, Lawrence; Gallo, Jason; Hertzfeld, Henry; Jolly, Claire; Macauley, Molly K.; Shapiro, Carl; Smart, Alan

    2016-03-10

    The production and accessibility of geospatial information including Earth observation is changing greatly both technically and in terms of human participation. Advances in technology have changed the way that geospatial data are produced and accessed, resulting in more efficient processes and greater accessibility than ever before. Improved technology has also created opportunities for increased participation in the gathering and interpretation of data through crowdsourcing and citizen science efforts. Increased accessibility has resulted in greater participation in the use of data as prices for Government-produced data have fallen and barriers to access have been reduced.

  20. Sextant: Visualizing time-evolving linked geospatial data

    NARCIS (Netherlands)

    C. Nikolaou (Charalampos); K. Dogani (Kallirroi); K. Bereta (Konstantina); G. Garbis (George); M. Karpathiotakis (Manos); K. Kyzirakos (Konstantinos); M. Koubarakis (Manolis)

    2015-01-01

    textabstractThe linked open data cloud is constantly evolving as datasets get continuously updated with newer versions. As a result, representing, querying, and visualizing the temporal dimension of linked data is crucial. This is especially important for geospatial datasets that form the backbone

  1. Geospatial metadata retrieval from web services

    Directory of Open Access Journals (Sweden)

    Ivanildo Barbosa

    Full Text Available Nowadays, producers of geospatial data in either raster or vector formats are able to make them available on the World Wide Web by deploying web services that enable users to access and query on those contents even without specific software for geoprocessing. Several providers around the world have deployed instances of WMS (Web Map Service, WFS (Web Feature Service and WCS (Web Coverage Service, all of them specified by the Open Geospatial Consortium (OGC. In consequence, metadata about the available contents can be retrieved to be compared with similar offline datasets from other sources. This paper presents a brief summary and describes the matching process between the specifications for OGC web services (WMS, WFS and WCS and the specifications for metadata required by the ISO 19115 - adopted as reference for several national metadata profiles, including the Brazilian one. This process focuses on retrieving metadata about the identification and data quality packages as well as indicates the directions to retrieve metadata related to other packages. Therefore, users are able to assess whether the provided contents fit to their purposes.

  2. A cross-sectional ecological analysis of international and sub-national health inequalities in commercial geospatial resource availability.

    Science.gov (United States)

    Dotse-Gborgbortsi, Winfred; Wardrop, Nicola; Adewole, Ademola; Thomas, Mair L H; Wright, Jim

    2018-05-23

    Commercial geospatial data resources are frequently used to understand healthcare utilisation. Although there is widespread evidence of a digital divide for other digital resources and infra-structure, it is unclear how commercial geospatial data resources are distributed relative to health need. To examine the distribution of commercial geospatial data resources relative to health needs, we assembled coverage and quality metrics for commercial geocoding, neighbourhood characterisation, and travel time calculation resources for 183 countries. We developed a country-level, composite index of commercial geospatial data quality/availability and examined its distribution relative to age-standardised all-cause and cause specific (for three main causes of death) mortality using two inequality metrics, the slope index of inequality and relative concentration index. In two sub-national case studies, we also examined geocoding success rates versus area deprivation by district in Eastern Region, Ghana and Lagos State, Nigeria. Internationally, commercial geospatial data resources were inversely related to all-cause mortality. This relationship was more pronounced when examining mortality due to communicable diseases. Commercial geospatial data resources for calculating patient travel times were more equitably distributed relative to health need than resources for characterising neighbourhoods or geocoding patient addresses. Countries such as South Africa have comparatively high commercial geospatial data availability despite high mortality, whilst countries such as South Korea have comparatively low data availability and low mortality. Sub-nationally, evidence was mixed as to whether geocoding success was lowest in more deprived districts. To our knowledge, this is the first global analysis of commercial geospatial data resources in relation to health outcomes. In countries such as South Africa where there is high mortality but also comparatively rich commercial geospatial

  3. Research and Practical Trends in Geospatial Sciences

    Science.gov (United States)

    Karpik, A. P.; Musikhin, I. A.

    2016-06-01

    In recent years professional societies have been undergoing fundamental restructuring brought on by extensive technological change and rapid evolution of geospatial science. Almost all professional communities have been affected. Communities are embracing digital techniques, modern equipment, software and new technological solutions at a staggering pace. In this situation, when planning financial investments and intellectual resource management, it is crucial to have a clear understanding of those trends that will be in great demand in 3-7 years. This paper reviews current scientific and practical activities of such non-governmental international organizations as International Federation of Surveyors, International Cartographic Association, and International Society for Photogrammetry and Remote Sensing, analyzes and groups most relevant topics brought up at their scientific events, forecasts most probable research and practical trends in geospatial sciences, outlines topmost leading countries and emerging markets for further detailed analysis of their activities, types of scientific cooperation and joint implementation projects.

  4. Imprementation of Vgi-Based Geoportal for Empowering Citizen's Geospatial Observatories Related to Urban Disaster Management

    Science.gov (United States)

    Lee, Sanghoon

    2016-06-01

    The volunteered geospatial information (VGI) will be efficient and cost-effective method for generating and sharing large disasterrelated geospatial data. The national mapping organizations have played the role of major geospatial collector have been moving toward the considering public participation data collecting method. Due to VGI can conduct to encourage public participation and empower citizens, mapping agency could make a partnership with members of the VGI community to help to provide well-structured geospatial data. In order to effectively be understood and sharing the public semantics, datasets and action model of the public participation GeoPortal, the implemented VGI-GeoPortal designated as the basis of ISO 19154, ISO 19101 and OGC Reference Model. The proof of concepts of VGI-GeoPortal has been implemented urban flooding use-case in Republic of Korea to collect from the public, and analyze disaster-related geospatial data including high-disaster potential information such as the location of poor drainage sewer, small signs of occurring landslide, flooding vulnerability of urban structure, and etc.

  5. Streamlining geospatial metadata in the Semantic Web

    Science.gov (United States)

    Fugazza, Cristiano; Pepe, Monica; Oggioni, Alessandro; Tagliolato, Paolo; Carrara, Paola

    2016-04-01

    In the geospatial realm, data annotation and discovery rely on a number of ad-hoc formats and protocols. These have been created to enable domain-specific use cases generalized search is not feasible for. Metadata are at the heart of the discovery process and nevertheless they are often neglected or encoded in formats that either are not aimed at efficient retrieval of resources or are plainly outdated. Particularly, the quantum leap represented by the Linked Open Data (LOD) movement did not induce so far a consistent, interlinked baseline in the geospatial domain. In a nutshell, datasets, scientific literature related to them, and ultimately the researchers behind these products are only loosely connected; the corresponding metadata intelligible only to humans, duplicated on different systems, seldom consistently. Instead, our workflow for metadata management envisages i) editing via customizable web- based forms, ii) encoding of records in any XML application profile, iii) translation into RDF (involving the semantic lift of metadata records), and finally iv) storage of the metadata as RDF and back-translation into the original XML format with added semantics-aware features. Phase iii) hinges on relating resource metadata to RDF data structures that represent keywords from code lists and controlled vocabularies, toponyms, researchers, institutes, and virtually any description one can retrieve (or directly publish) in the LOD Cloud. In the context of a distributed Spatial Data Infrastructure (SDI) built on free and open-source software, we detail phases iii) and iv) of our workflow for the semantics-aware management of geospatial metadata.

  6. Development of Geospatial Map Based Election Portal

    Science.gov (United States)

    Gupta, A. Kumar Chandra; Kumar, P.; Vasanth Kumar, N.

    2014-11-01

    The Geospatial Delhi Limited (GSDL), a Govt. of NCT of Delhi Company formed in order to provide the geospatial information of National Capital Territory of Delhi (NCTD) to the Government of National Capital Territory of Delhi (GNCTD) and its organs such as DDA, MCD, DJB, State Election Department, DMRC etc., for the benefit of all citizens of Government of National Capital Territory of Delhi (GNCTD). This paper describes the development of Geospatial Map based Election portal (GMEP) of NCT of Delhi. The portal has been developed as a map based spatial decision support system (SDSS) for pertain to planning and management of Department of Chief Electoral Officer, and as an election related information searching tools (Polling Station, Assembly and parliamentary constituency etc.,) for the citizens of NCTD. The GMEP is based on Client-Server architecture model. It has been developed using ArcGIS Server 10.0 with J2EE front-end on Microsoft Windows environment. The GMEP is scalable to enterprise SDSS with enterprise Geo Database & Virtual Private Network (VPN) connectivity. Spatial data to GMEP includes delimited precinct area boundaries of Voters Area of Polling stations, Assembly Constituency, Parliamentary Constituency, Election District, Landmark locations of Polling Stations & basic amenities (Police Stations, Hospitals, Schools and Fire Stations etc.). GMEP could help achieve not only the desired transparency and easiness in planning process but also facilitates through efficient & effective tools for management of elections. It enables a faster response to the changing ground realities in the development planning, owing to its in-built scientific approach and open-ended design.

  7. Securing the AliEn File Catalogue - Enforcing authorization with accountable file operations

    International Nuclear Information System (INIS)

    Schreiner, Steffen; Banerjee, Subho Sankar; Betev, Latchezar; Carminati, Federico; Vladimirovna Datskova, Olga; Furano, Fabrizio; Grigoras, Alina; Grigoras, Costin; Mendez Lorenzo, Patricia; Peters, Andreas Joachim; Saiz, Pablo; Bagnasco, Stefano; Zhu Jianlin

    2011-01-01

    The AliEn Grid Services, as operated by the ALICE Collaboration in its global physics analysis grid framework, is based on a central File Catalogue together with a distributed set of storage systems and the possibility to register links to external data resources. This paper describes several identified vulnerabilities in the AliEn File Catalogue access protocol regarding fraud and unauthorized file alteration and presents a more secure and revised design: a new mechanism, called LFN Booking Table, is introduced in order to keep track of access authorization in the transient state of files entering or leaving the File Catalogue. Due to a simplification of the original Access Envelope mechanism for xrootd-protocol-based storage systems, fundamental computational improvements of the mechanism were achieved as well as an up to 50% reduction of the credential's size. By extending the access protocol with signed status messages from the underlying storage system, the File Catalogue receives trusted information about a file's size and checksum and the protocol is no longer dependent on client trust. Altogether, the revised design complies with atomic and consistent transactions and allows for accountable, authentic, and traceable file operations. This paper describes these changes as part and beyond the development of AliEn version 2.19.

  8. AGWA: The Automated Geospatial Watershed Assessment Tool

    Science.gov (United States)

    The Automated Geospatial Watershed Assessment Tool (AGWA, see: www.tucson.ars.ag.gov/agwa or http://www.epa.gov/esd/land-sci/agwa/) is a GIS interface jointly developed by the USDA-Agricultural Research Service, the U.S. Environmental Protection Agency, the University of Arizona...

  9. Users Manual for the Geospatial Stream Flow Model (GeoSFM)

    Science.gov (United States)

    Artan, Guleid A.; Asante, Kwabena; Smith, Jodie; Pervez, Md Shahriar; Entenmann, Debbie; Verdin, James P.; Rowland, James

    2008-01-01

    The monitoring of wide-area hydrologic events requires the manipulation of large amounts of geospatial and time series data into concise information products that characterize the location and magnitude of the event. To perform these manipulations, scientists at the U.S. Geological Survey Center for Earth Resources Observation and Science (EROS), with the cooperation of the U.S. Agency for International Development, Office of Foreign Disaster Assistance (USAID/OFDA), have implemented a hydrologic modeling system. The system includes a data assimilation component to generate data for a Geospatial Stream Flow Model (GeoSFM) that can be run operationally to identify and map wide-area streamflow anomalies. GeoSFM integrates a geographical information system (GIS) for geospatial preprocessing and postprocessing tasks and hydrologic modeling routines implemented as dynamically linked libraries (DLLs) for time series manipulations. Model results include maps that depicting the status of streamflow and soil water conditions. This Users Manual provides step-by-step instructions for running the model and for downloading and processing the input data required for initial model parameterization and daily operation.

  10. VizieR Online Data Catalog: WATCH Solar X-Ray Burst Catalogue (Crosby+ 1998)

    Science.gov (United States)

    Crosby, N.; Lund, N.; Vilmer, N.; Sunyaev, R.

    1998-01-01

    Catalogue containing solar X-ray bursts measured by the Danish Wide Angle Telescope for Cosmic Hard X-Rays (WATCH) experiment aboard the Russian satellite GRANAT in the deca-keV energy range. Table 1 lists the periods during which solar observations with WATCH are available (WATCH ON-TIME) and where the bursts listed in the catalogue have been observed. (2 data files).

  11. Prototype-based analysis of GAMA galaxy catalogue data

    NARCIS (Netherlands)

    Nolte, A.; Wang, L.; Biehl, M; Verleysen, Michel

    2018-01-01

    We present a prototype-based machine learning analysis of labeled galaxy catalogue data containing parameters from the Galaxy and Mass Assembly (GAMA) survey. Using both an unsupervised and supervised method, the Self-Organizing Map and Generalized Relevance Matrix Learning Vec- tor Quantization, we

  12. Procedures and challenges of retrospective catalogue conversion in ...

    African Journals Online (AJOL)

    The study recommended that management of the universities should provide stand- by electricity generator and upgrading of Internet network services among other things in the two university libraries for effective and efficient service delivery. Key words: Catalogue, Libraries Procedures, Conversion, Universities ...

  13. The geospatial web how geobrowsers, social software and the web 2 0 are shaping the network society

    CERN Document Server

    Scharl, Arno; Tochtermann, Klaus

    2007-01-01

    The Geospatial Web will have a profound impact on managing knowledge, structuring work flows within and across organizations, and communicating with like-minded individuals in virtual communities. The enabling technologies for the Geospatial Web are geo-browsers such as NASA World Wind, Google Earth and Microsoft Live Local 3D. These three-dimensional platforms revolutionize the production and consumption of media products. They not only reveal the geographic distribution of Web resources and services, but also bring together people of similar interests, browsing behavior, or geographic location. This book summarizes the latest research on the Geospatial Web's technical foundations, describes information services and collaborative tools built on top of geo-browsers, and investigates the environmental, social and economic impacts of geospatial applications. The role of contextual knowledge in shaping the emerging network society deserves particular attention. By integrating geospatial and semantic technology, ...

  14. Identification of stars and digital version of the catalogue of 1958 by Brodskaya and Shajn

    Science.gov (United States)

    Gorbunov, M. A.; Shlyapnikov, A. A.

    2017-12-01

    The following topics are considered: the identification of objects on search maps, the determination of their coordinates at the epoch of 2000, and converting the published version of the catalogue of 1958 by Brodskaya and Shajn into a machine-readable format. The statistics for photometric and spectral data from the original catalogue is presented. A digital version of the catalogue is described, as well as its presentation in HTML, VOTable and AJS formats and the basic principles of work in the interactive application of International Virtual Observatory - the Aladin Sky Atlas.

  15. A study on state of Geospatial courses in Indian Universities

    Science.gov (United States)

    Shekhar, S.

    2014-12-01

    Today the world is dominated by three technologies such as Nano technology, Bio technology and Geospatial technology. This increases the huge demand for experts in the respective field for disseminating the knowledge as well as for an innovative research. Therefore, the prime need is to train the existing fraternity to gain progressive knowledge in these technologies and impart the same to student community. The geospatial technology faces some peculiar problem than other two technologies because of its interdisciplinary, multi-disciplinary nature. It attracts students and mid career professionals from various disciplines including Physics, Computer science, Engineering, Geography, Geology, Agriculture, Forestry, Town Planning and so on. Hence there is always competition to crab and stabilize their position. The students of Master's degree in Geospatial science are facing two types of problem. The first one is no unique identity in the academic field. Neither they are exempted for National eligibility Test for Lecturer ship nor given an opportunity to have the exam in geospatial science. The second one is differential treatment by the industrial world. The students are either given low grade jobs or poorly paid for their job. Thus, it is a serious issue about the future of this course in the Universities and its recognition in the academic and industrial world. The universities should make this course towards more job oriented in consultation with the Industries and Industries should come forward to share their demands and requirements to the Universities, so that necessary changes in the curriculum can be made to meet the industrial requirements.

  16. A Metadata Schema for Geospatial Resource Discovery Use Cases

    Directory of Open Access Journals (Sweden)

    Darren Hardy

    2014-07-01

    Full Text Available We introduce a metadata schema that focuses on GIS discovery use cases for patrons in a research library setting. Text search, faceted refinement, and spatial search and relevancy are among GeoBlacklight's primary use cases for federated geospatial holdings. The schema supports a variety of GIS data types and enables contextual, collection-oriented discovery applications as well as traditional portal applications. One key limitation of GIS resource discovery is the general lack of normative metadata practices, which has led to a proliferation of metadata schemas and duplicate records. The ISO 19115/19139 and FGDC standards specify metadata formats, but are intricate, lengthy, and not focused on discovery. Moreover, they require sophisticated authoring environments and cataloging expertise. Geographic metadata standards target preservation and quality measure use cases, but they do not provide for simple inter-institutional sharing of metadata for discovery use cases. To this end, our schema reuses elements from Dublin Core and GeoRSS to leverage their normative semantics, community best practices, open-source software implementations, and extensive examples already deployed in discovery contexts such as web search and mapping. Finally, we discuss a Solr implementation of the schema using a "geo" extension to MODS.

  17. RESEARCH AND PRACTICAL TRENDS IN GEOSPATIAL SCIENCES

    Directory of Open Access Journals (Sweden)

    A. P. Karpik

    2016-06-01

    Full Text Available In recent years professional societies have been undergoing fundamental restructuring brought on by extensive technological change and rapid evolution of geospatial science. Almost all professional communities have been affected. Communities are embracing digital techniques, modern equipment, software and new technological solutions at a staggering pace. In this situation, when planning financial investments and intellectual resource management, it is crucial to have a clear understanding of those trends that will be in great demand in 3-7 years. This paper reviews current scientific and practical activities of such non-governmental international organizations as International Federation of Surveyors, International Cartographic Association, and International Society for Photogrammetry and Remote Sensing, analyzes and groups most relevant topics brought up at their scientific events, forecasts most probable research and practical trends in geospatial sciences, outlines topmost leading countries and emerging markets for further detailed analysis of their activities, types of scientific cooperation and joint implementation projects.

  18. Multiband Study of Radio Sources of the RCR Catalogue with Virtual Observatory Tools

    Directory of Open Access Journals (Sweden)

    Zhelenkova O. P.

    2012-09-01

    Full Text Available We present early results of our multiband study of the RATAN Cold Revised (RCR catalogue obtained from seven cycles of the “Cold” survey carried with the RATAN-600 radio telescope at 7.6 cm in 1980-1999, at the declination of the SS 433 source. We used the 2MASS and LAS UKIDSS infrared surveys, the DSS-II and SDSS DR7 optical surveys, as well as the USNO-B1 and GSC-II catalogues, the VLSS, TXS, NVSS, FIRST and GB6 radio surveys to accumulate information about the sources. For radio sources that have no detectable optical candidate in optical or infrared catalogues, we additionally looked through images in several bands from the SDSS, LAS UKIDSS, DPOSS, 2MASS surveys and also used co-added frames in different bands. We reliably identified 76% of radio sources of the RCR catalogue. We used the ALADIN and SAOImage DS9 scripting capabilities, interoperability services of ALADIN and TOPCAT, and also other Virtual Observatory (VO tools and resources, such as CASJobs, NED, Vizier, and WSA, for effective data access, visualization and analysis. Without VO tools it would have been problematic to perform our study.

  19. Geospatial technology perspectives for mining vis-a-vis sustainable forest ecosystems

    Directory of Open Access Journals (Sweden)

    Goparaju Laxmi

    2017-06-01

    Full Text Available Forests, the backbone of biogeochemical cycles and life supporting systems, are under severe pressure due to varied anthropogenic activities. Mining activities are one among the major reasons for forest destruction questioning the survivability and sustainability of flora and fauna existing in that area. Thus, monitoring and managing the impact of mining activities on natural resources at regular intervals is necessary to check the status of their depleted conditions, and to take up restoration and conservative measurements. Geospatial technology provides means to identify the impact of different mining operations on forest ecosystems and helps in proposing initiatives for safeguarding the forest environment. In this context, the present study highlights the problems related to mining in forest ecosystems and elucidates how geospatial technology can be employed at various stages of mining activities to achieve a sustainable forest ecosystem. The study collates information from various sources and highlights the role of geospatial technology in mining industries and reclamation process.

  20. VizieR Online Data Catalog: Catalogue of Galactic Planetary Nebulae (Kohoutek, 2001)

    Science.gov (United States)

    Kohoutek, L.

    2001-05-01

    The "Catalogue of Galactic Planetary Nebulae (Version 2000)" appears in Abhandlungen aus der Hamburger Sternwarte, Band XII in the year 2001. It is a continuation of CGPN(1967) and contains 1510 objects classified as galactic PNe up to the end of 1999. The lists of possible pre-PNe and possible post-PNe are also given. The catalogue is restricted only to the data belonging to the location and identification of the objects. It gives identification charts of PNe discovered since 1965 (published in the supplements to CGPN) and those charts of objects discovered earlier, which have wrong or uncertain identification. The question "what is a planetary nebula" is discussed and the typical values of PNe and of their central stars are summarized. Short statistics about the discoveries of PNe are given. The catalogue is also available in the Centre de Donnees, Strasbourg and at Hamburg Observatory via internet. (15 data files).

  1. Solar Maps | Geospatial Data Science | NREL

    Science.gov (United States)

    Solar Maps Solar Maps These solar maps provide average daily total solar resource information on disability, contact the Geospatial Data Science Team. U.S. State Solar Resource Maps Access state maps of MT NE NV NH NJ NM NY NC ND OH OK OR PA RI SC SD TN TX UT VT VA WA WV WI WY × U.S. Solar Resource

  2. Distributed Storage Algorithm for Geospatial Image Data Based on Data Access Patterns.

    Directory of Open Access Journals (Sweden)

    Shaoming Pan

    Full Text Available Declustering techniques are widely used in distributed environments to reduce query response time through parallel I/O by splitting large files into several small blocks and then distributing those blocks among multiple storage nodes. Unfortunately, however, many small geospatial image data files cannot be further split for distributed storage. In this paper, we propose a complete theoretical system for the distributed storage of small geospatial image data files based on mining the access patterns of geospatial image data using their historical access log information. First, an algorithm is developed to construct an access correlation matrix based on the analysis of the log information, which reveals the patterns of access to the geospatial image data. Then, a practical heuristic algorithm is developed to determine a reasonable solution based on the access correlation matrix. Finally, a number of comparative experiments are presented, demonstrating that our algorithm displays a higher total parallel access probability than those of other algorithms by approximately 10-15% and that the performance can be further improved by more than 20% by simultaneously applying a copy storage strategy. These experiments show that the algorithm can be applied in distributed environments to help realize parallel I/O and thereby improve system performance.

  3. Distributed Storage Algorithm for Geospatial Image Data Based on Data Access Patterns.

    Science.gov (United States)

    Pan, Shaoming; Li, Yongkai; Xu, Zhengquan; Chong, Yanwen

    2015-01-01

    Declustering techniques are widely used in distributed environments to reduce query response time through parallel I/O by splitting large files into several small blocks and then distributing those blocks among multiple storage nodes. Unfortunately, however, many small geospatial image data files cannot be further split for distributed storage. In this paper, we propose a complete theoretical system for the distributed storage of small geospatial image data files based on mining the access patterns of geospatial image data using their historical access log information. First, an algorithm is developed to construct an access correlation matrix based on the analysis of the log information, which reveals the patterns of access to the geospatial image data. Then, a practical heuristic algorithm is developed to determine a reasonable solution based on the access correlation matrix. Finally, a number of comparative experiments are presented, demonstrating that our algorithm displays a higher total parallel access probability than those of other algorithms by approximately 10-15% and that the performance can be further improved by more than 20% by simultaneously applying a copy storage strategy. These experiments show that the algorithm can be applied in distributed environments to help realize parallel I/O and thereby improve system performance.

  4. A Spatial Data Infrastructure to Share Earth and Space Science Data

    Science.gov (United States)

    Nativi, S.; Mazzetti, P.; Bigagli, L.; Cuomo, V.

    2006-05-01

    Spatial Data Infrastructure:SDI (also known as Geospatial Data Infrastructure) is fundamentally a mechanism to facilitate the sharing and exchange of geospatial data. SDI is a scheme necessary for the effective collection, management, access, delivery and utilization of geospatial data; it is important for: objective decision making and sound land based policy, support economic development and encourage socially and environmentally sustainable development. As far as data model and semantics are concerned, a valuable and effective SDI should be able to cross the boundaries between the Geographic Information System/Science (GIS) and Earth and Space Science (ESS) communities. Hence, SDI should be able to discover, access and share information and data produced and managed by both GIS and ESS communities, in an integrated way. In other terms, SDI must be built on a conceptual and technological framework which abstracts the nature and structure of shared dataset: feature-based data or Imagery, Gridded and Coverage Data (IGCD). ISO TC211 and the Open Geospatial Consortium provided important artifacts to build up this framework. In particular, the OGC Web Services (OWS) initiatives and several Interoperability Experiment (e.g. the GALEON IE) are extremely useful for this purpose. We present a SDI solution which is able to manage both GIS and ESS datasets. It is based on OWS and other well-accepted or promising technologies, such as: UNIDATA netCDF and CDM, ncML and ncML-GML. Moreover, it uses a specific technology to implement a distributed and federated system of catalogues: the GI-Cat. This technology performs data model mediation and protocol adaptation tasks. It is used to work out a metadata clearinghouse service, implementing a common (federal) catalogue model which is based on the ISO 19115 core metadata for geo-dataset. Nevertheless, other well- accepted or standard catalogue data models can be easily implemented as common view (e.g. OGC CS-W, the next coming

  5. The Academic SDI—Towards understanding spatial data infrastructures for research and education

    CSIR Research Space (South Africa)

    Coetzee, S

    2017-05-01

    Full Text Available facilitating and coordinating the exchange of geospatial data and services between stakeholders from different levels in the spatial data community. Universities and other research organisations typically have well-established libraries and digital catalogues...

  6. Remote Sensing Technologies and Geospatial Modelling Hierarchy for Smart City Support

    Science.gov (United States)

    Popov, M.; Fedorovsky, O.; Stankevich, S.; Filipovich, V.; Khyzhniak, A.; Piestova, I.; Lubskyi, M.; Svideniuk, M.

    2017-12-01

    The approach to implementing the remote sensing technologies and geospatial modelling for smart city support is presented. The hierarchical structure and basic components of the smart city information support subsystem are considered. Some of the already available useful practical developments are described. These include city land use planning, urban vegetation analysis, thermal condition forecasting, geohazard detection, flooding risk assessment. Remote sensing data fusion approach for comprehensive geospatial analysis is discussed. Long-term city development forecasting by Forrester - Graham system dynamics model is provided over Kiev urban area.

  7. Architecture of a Process Broker for Interoperable Geospatial Modeling on the Web

    Directory of Open Access Journals (Sweden)

    Lorenzo Bigagli

    2015-04-01

    Full Text Available The identification of appropriate mechanisms for process sharing and reuse by means of composition is considered a key enabler for the effective uptake of a global Earth Observation infrastructure, currently pursued by the international geospatial research community. Modelers in need of running complex workflows may benefit from outsourcing process composition to a dedicated external service, according to the brokering approach. This work introduces our architecture of a process broker, as a distributed information system for creating, validating, editing, storing, publishing and executing geospatial-modeling workflows. The broker provides a service framework for adaptation, reuse and complementation of existing processing resources (including models and geospatial services in general in the form of interoperable, executable workflows. The described solution has been experimentally applied in several use scenarios in the context of EU-funded projects and the Global Earth Observation System of Systems.

  8. Partially populated catalogue of measured properties of field sections.

    Science.gov (United States)

    2014-10-01

    This catalogue documents the construction, monitoring, and mixture information of 11 test sections: four in SH 15 in the north Amarillo, three in US 62 in Childress, and four in Loop 820 in Fort Worth.

  9. Halo substructure in the SDSS-Gaia catalogue: streams and clumps

    Science.gov (United States)

    Myeong, G. C.; Evans, N. W.; Belokurov, V.; Amorisco, N. C.; Koposov, S. E.

    2018-04-01

    We use the Sloan Digital Sky Survey (SDSS)-Gaia Catalogue to identify six new pieces of halo substructure. SDSS-Gaia is an astrometric catalogue that exploits SDSS data release 9 to provide first epoch photometry for objects in the Gaia source catalogue. We use a version of the catalogue containing 245 316 stars with all phase-space coordinates within a heliocentric distance of ˜10 kpc. We devise a method to assess the significance of halo substructures based on their clustering in velocity space. The two most substantial structures are multiple wraps of a stream which has undergone considerable phase mixing (S1, with 94 members) and a kinematically cold stream (S2, with 61 members). The member stars of S1 have a median position of (X, Y, Z) = (8.12, -0.22, 2.75) kpc and a median metallicity of [Fe/H] = -1.78. The stars of S2 have median coordinates (X, Y, Z) = (8.66, 0.30, 0.77) kpc and a median metallicity of [Fe/H] = -1.91. They lie in velocity space close to some of the stars in the stream reported by Helmi et al. By modelling, we estimate that both structures had progenitors with virial masses ≈1010M⊙ and infall times ≳ 9 Gyr ago. Using abundance matching, these correspond to stellar masses between 106 and 107M⊙. These are somewhat larger than the masses inferred through the mass-metallicity relation by factors of 5 to 15. Additionally, we identify two further substructures (S3 and S4 with 55 and 40 members) and two clusters or moving group (C1 and C2 with 24 and 12) members. In all six cases, clustering in kinematics is found to correspond to clustering in both configuration space and metallicity, adding credence to the reliability of our detections.

  10. Provisional host catalogue of Fig wasps (Hymenoptera, Chalcidoidea)

    NARCIS (Netherlands)

    Wiebes, J.T.

    1966-01-01

    INTRODUCTION In this catalogue — entitled "provisional" because our knowledge of the subject is still so evidently incomplete — all species of Ficus mentioned as hosts of fig wasps, are listed with the Hymenoptera Chalcidoidea reared from their receptacles. The names used for the Agaonidae are in

  11. White dwarf-main sequence binaries from LAMOST: the DR5 catalogue

    Science.gov (United States)

    Ren, J.-J.; Rebassa-Mansergas, A.; Parsons, S. G.; Liu, X.-W.; Luo, A.-L.; Kong, X.; Zhang, H.-T.

    2018-03-01

    We present the data release (DR) 5 catalogue of white dwarf-main sequence (WDMS) binaries from the Large Area Multi-Object fiber Spectroscopic Telescope (LAMOST). The catalogue contains 876 WDMS binaries, of which 757 are additions to our previous LAMOST DR1 sample and 357 are systems that have not been published before. We also describe a LAMOST-dedicated survey that aims at obtaining spectra of photometrically-selected WDMS binaries from the Sloan Digital Sky Survey (SDSS) that are expected to contain cool white dwarfs and/or early type M dwarf companions. This is a population under-represented in previous SDSS WDMS binary catalogues. We determine the stellar parameters (white dwarf effective temperatures, surface gravities and masses, and M dwarf spectral types) of the LAMOST DR5 WDMS binaries and make use of the parameter distributions to analyse the properties of the sample. We find that, despite our efforts, systems containing cool white dwarfs remain under-represented. Moreover, we make use of LAMOST DR5 and SDSS DR14 (when available) spectra to measure the Na I λλ 8183.27, 8194.81 absorption doublet and/or Hα emission radial velocities of our systems. This allows identifying 128 binaries displaying significant radial velocity variations, 76 of which are new. Finally, we cross-match our catalogue with the Catalina Surveys and identify 57 systems displaying light curve variations. These include 16 eclipsing systems, two of which are new, and nine binaries that are new eclipsing candidates. We calculate periodograms from the photometric data and measure (estimate) the orbital periods of 30 (15) WDMS binaries.

  12. IMPREMENTATION OF VGI-BASED GEOPORTAL FOR EMPOWERING CITIZEN’S GEOSPATIAL OBSERVATORIES RELATED TO URBAN DISASTER MANAGEMENT

    Directory of Open Access Journals (Sweden)

    S. Lee

    2016-06-01

    Full Text Available The volunteered geospatial information (VGI will be efficient and cost-effective method for generating and sharing large disasterrelated geospatial data. The national mapping organizations have played the role of major geospatial collector have been moving toward the considering public participation data collecting method. Due to VGI can conduct to encourage public participation and empower citizens, mapping agency could make a partnership with members of the VGI community to help to provide well-structured geospatial data. In order to effectively be understood and sharing the public semantics, datasets and action model of the public participation GeoPortal, the implemented VGI-GeoPortal designated as the basis of ISO 19154, ISO 19101 and OGC Reference Model. The proof of concepts of VGI-GeoPortal has been implemented urban flooding use-case in Republic of Korea to collect from the public, and analyze disaster-related geospatial data including high-disaster potential information such as the location of poor drainage sewer, small signs of occurring landslide, flooding vulnerability of urban structure, and etc.

  13. FUNGIBLE AND COMPATIBLE BIOFUELS: LITERATURE SEARCH, SUMMARY, AND RECOMMENDATIONS

    Energy Technology Data Exchange (ETDEWEB)

    Bunting, Bruce G [ORNL; Bunce, Michael [ORNL; Barone, Teresa L [ORNL; Storey, John Morse [ORNL

    2011-04-01

    The purpose of the study described in this report is to summarize the various barriers to more widespread distribution of bio-fuels through our common carrier fuel distribution system, which includes pipelines, barges and rail, fuel tankage, and distribution terminals. Addressing these barriers is necessary to allow the more widespread utilization and distribution of bio-fuels, in support of a renewable fuels standard and possible future low-carbon fuel standards. These barriers can be classified into several categories, including operating practice, regulatory, technical, and acceptability barriers. Possible solutions to these issues are discussed; including compatibility evaluation, changes to bio-fuels, regulatory changes, and changes in the distribution system or distribution practices. No actual experimental research has been conducted in the writing of this report, but results are used to develop recommendations for future research and additional study as appropriate. This project addresses recognized barriers to the wider use of bio-fuels in the areas of development of codes and standards, industrial and consumer awareness, and materials compatibility issues.

  14. Adoption of Geospatial Systems towards evolving Sustainable Himalayan Mountain Development

    Science.gov (United States)

    Murthy, M. S. R.; Bajracharya, B.; Pradhan, S.; Shestra, B.; Bajracharya, R.; Shakya, K.; Wesselmann, S.; Ali, M.; Bajracharya, S.; Pradhan, S.

    2014-11-01

    Natural resources dependence of mountain communities, rapid social and developmental changes, disaster proneness and climate change are conceived as the critical factors regulating sustainable Himalayan mountain development. The Himalayan region posed by typical geographic settings, diverse physical and cultural diversity present a formidable challenge to collect and manage data, information and understands varied socio-ecological settings. Recent advances in earth observation, near real-time data, in-situ measurements and in combination of information and communication technology have transformed the way we collect, process, and generate information and how we use such information for societal benefits. Glacier dynamics, land cover changes, disaster risk reduction systems, food security and ecosystem conservation are a few thematic areas where geospatial information and knowledge have significantly contributed to informed decision making systems over the region. The emergence and adoption of near-real time systems, unmanned aerial vehicles (UAV), board-scale citizen science (crowd-sourcing), mobile services and mapping, and cloud computing have paved the way towards developing automated environmental monitoring systems, enhanced scientific understanding of geophysical and biophysical processes, coupled management of socio-ecological systems and community based adaptation models tailored to mountain specific environment. There are differentiated capacities among the ICIMOD regional member countries with regard to utilization of earth observation and geospatial technologies. The region can greatly benefit from a coordinated and collaborative approach to capture the opportunities offered by earth observation and geospatial technologies. The regional level data sharing, knowledge exchange, and Himalayan GEO supporting geospatial platforms, spatial data infrastructure, unique region specific satellite systems to address trans-boundary challenges would go a long way in

  15. Advancing Collaborative Climate Studies through Globally Distributed Geospatial Analysis

    Science.gov (United States)

    Singh, R.; Percivall, G.

    2009-12-01

    Infrastructure and the broader GEOSS architecture. Of specific interest to this session is the work on geospatial workflows and geo-processing and data discovery and access. CCIP demonstrates standards-based interoperability between geospatial applications in the service of Climate Change analysis. CCIP is planned to be a yearly exercise. It consists of a network of online data services (WCS, WFS, SOS), analysis services (WPS, WCPS, WMS), and clients that exercise those services. In 2009, CCIP focuses on Australia, and the initial application of existing OGC services to climate studies. The results of the 2009 CCIP will serve as requirements for more complex geo-processing services to be developed for CCIP 2010. The benefits of CCIP include accelerating the implementation of the GCOS, and building confidence that implementations using multi-vendor interoperable technologies can help resolve vexing climate change questions. AIP-2: Architecture Implementation Pilot, Phase 2 CCIP: Climate Challenge Integration Plugfest GEO: Group on Earth Observations GEOSS: Global Earth Observing System of Systems GCOS: Global Climate Observing System OGC: Open Geospatial Consortium SOS: Sensor Observation Service WCS: Web Coverage Service WCPS: Web Coverage Processing Service WFS: Web Feature Service WMS: Web Mapping Service

  16. Transportation of Large Wind Components: A Review of Existing Geospatial Data

    Energy Technology Data Exchange (ETDEWEB)

    Mooney, Meghan [National Renewable Energy Lab. (NREL), Golden, CO (United States); Maclaurin, Galen [National Renewable Energy Lab. (NREL), Golden, CO (United States)

    2016-09-01

    This report features the geospatial data component of a larger project evaluating logistical and infrastructure requirements for transporting oversized and overweight (OSOW) wind components. The goal of the larger project was to assess the status and opportunities for improving the infrastructure and regulatory practices necessary to transport wind turbine towers, blades, and nacelles from current and potential manufacturing facilities to end-use markets. The purpose of this report is to summarize existing geospatial data on wind component transportation infrastructure and to provide a data gap analysis, identifying areas for further analysis and data collection.

  17. A catalogue of the genera of the Vespidae (Hymenoptera)

    NARCIS (Netherlands)

    Vecht, van der J.; Carpenter, J.M.

    1990-01-01

    A comprehensive generic catalogue of the Vespidae is presented. New nomenclatural changes include synonymy of Alastoroides Saussure, 1856, with Paralastor Saussure 1856; Araucodynerus Willink, 1968, with Hypodynerus Saussure 1855; and Paranortonia Bequaert, 1940, with Parazumia Saussure, 1855.

  18. INTEGRATING GEOSPATIAL TECHNOLOGIES AND SECONDARY STUDENT PROJECTS: THE GEOSPATIAL SEMESTER

    Directory of Open Access Journals (Sweden)

    Bob Kolvoord

    2012-12-01

    Full Text Available Resumen:El Semestre Geoespacial es una actividad de educación geográfica centrada en que los estudiantes del último curso de secundaria en los institutos norteamericanos, adquieran competencias y habilidades específicas en sistemas de información geográfica, GPS y teledetección. A través de una metodología de aprendizaje basado en proyectos, los alumnos se motivan e implican en la realización de trabajos de investigación en los que analizan, e incluso proponen soluciones, diferentes procesos, problemas o cuestiones de naturaleza espacial. El proyecto está coordinado por la Universidad James Madison y lleva siete años implantándose en diferentes institutos del Estado de Virginia, implicando a más de 20 centros educativos y 1.500 alumnos. Los alumnos que superan esta asignatura de la enseñanza secundaria obtienen la convalidación de determinados créditos académicos de la Universidad de referencia.Palabras clave:Sistemas de información geográfica, enseñanza, didáctica de la geografía, semestre geoespacial.Abstract:The Geospatial Semester is a geographical education activity focused on students in their final year of secondary schools in the U.S., acquiring specific skills in GIS, GPS and remote sensing. Through a methodology for project-based learning, students are motivated and involved in conducting research using geographic information systems and analyze, and even propose solutions, different processes, problems or issues spatial in nature. The Geospatial Semester university management not only ensures proper coaching, guidance and GIS training for teachers of colleges, but has established a system whereby students who pass this course of secondary education gain the recognition of certain credits from the University.Key words:Geographic information system, teaching, geographic education, geospatial semester. Résumé:Le semestre géospatial est une activité axée sur l'éducation géographique des étudiants en derni

  19. Automatic Scaling Hadoop in the Cloud for Efficient Process of Big Geospatial Data

    Directory of Open Access Journals (Sweden)

    Zhenlong Li

    2016-09-01

    Full Text Available Efficient processing of big geospatial data is crucial for tackling global and regional challenges such as climate change and natural disasters, but it is challenging not only due to the massive data volume but also due to the intrinsic complexity and high dimensions of the geospatial datasets. While traditional computing infrastructure does not scale well with the rapidly increasing data volume, Hadoop has attracted increasing attention in geoscience communities for handling big geospatial data. Recently, many studies were carried out to investigate adopting Hadoop for processing big geospatial data, but how to adjust the computing resources to efficiently handle the dynamic geoprocessing workload was barely explored. To bridge this gap, we propose a novel framework to automatically scale the Hadoop cluster in the cloud environment to allocate the right amount of computing resources based on the dynamic geoprocessing workload. The framework and auto-scaling algorithms are introduced, and a prototype system was developed to demonstrate the feasibility and efficiency of the proposed scaling mechanism using Digital Elevation Model (DEM interpolation as an example. Experimental results show that this auto-scaling framework could (1 significantly reduce the computing resource utilization (by 80% in our example while delivering similar performance as a full-powered cluster; and (2 effectively handle the spike processing workload by automatically increasing the computing resources to ensure the processing is finished within an acceptable time. Such an auto-scaling approach provides a valuable reference to optimize the performance of geospatial applications to address data- and computational-intensity challenges in GIScience in a more cost-efficient manner.

  20. An Automated End-To Multi-Agent Qos Based Architecture for Selection of Geospatial Web Services

    Science.gov (United States)

    Shah, M.; Verma, Y.; Nandakumar, R.

    2012-07-01

    Over the past decade, Service-Oriented Architecture (SOA) and Web services have gained wide popularity and acceptance from researchers and industries all over the world. SOA makes it easy to build business applications with common services, and it provides like: reduced integration expense, better asset reuse, higher business agility, and reduction of business risk. Building of framework for acquiring useful geospatial information for potential users is a crucial problem faced by the GIS domain. Geospatial Web services solve this problem. With the help of web service technology, geospatial web services can provide useful geospatial information to potential users in a better way than traditional geographic information system (GIS). A geospatial Web service is a modular application designed to enable the discovery, access, and chaining of geospatial information and services across the web that are often both computation and data-intensive that involve diverse sources of data and complex processing functions. With the proliferation of web services published over the internet, multiple web services may provide similar functionality, but with different non-functional properties. Thus, Quality of Service (QoS) offers a metric to differentiate the services and their service providers. In a quality-driven selection of web services, it is important to consider non-functional properties of the web service so as to satisfy the constraints or requirements of the end users. The main intent of this paper is to build an automated end-to-end multi-agent based solution to provide the best-fit web service to service requester based on QoS.

  1. 75 FR 54546 - Amendment of the Commission's Rules Governing Hearing Aid-Compatible Mobile Handsets

    Science.gov (United States)

    2010-09-08

    ... compatibility standards are already being met for handsets that operate on a variety of 2G and 3G air interfaces... the Commission's Rules Governing Hearing Aid- Compatible Mobile Handsets AGENCY: Federal... to extend the scope of the rules beyond the current category of Commercial Mobile Radio Service (CMRS...

  2. A global catalogue of Ceres impact craters ≥ 1 km and preliminary analysis

    Science.gov (United States)

    Gou, Sheng; Yue, Zongyu; Di, Kaichang; Liu, Zhaoqin

    2018-03-01

    The orbital data products of Ceres, including global LAMO image mosaic and global HAMO DTM with a resolution of 35 m/pixel and 135 m/pixel respectively, are utilized in this research to create a global catalogue of impact craters with diameter ≥ 1 km, and their morphometric parameters are calculated. Statistics shows: (1) There are 29,219 craters in the catalogue, and the craters have a various morphologies, e.g., polygonal crater, floor fractured crater, complex crater with central peak, etc.; (2) The identifiable smallest crater size is extended to 1 km and the crater numbers have been updated when compared with the crater catalogue (D ≥ 20 km) released by the Dawn Science Team; (3) The d/D ratios for fresh simple craters, obviously degraded simple crater and polygonal simple crater are 0.11 ± 0.04, 0.05 ± 0.04 and 0.14 ± 0.02 respectively. (4) The d/D ratios for non-polygonal complex crater and polygonal complex crater are 0.08 ± 0.04 and 0.09 ± 0.03. The global crater catalogue created in this work can be further applied to many other scientific researches, such as comparing d/D with other bodies, inferring subsurface properties, determining surface age, and estimating average erosion rate.

  3. The PMA Catalogue as a realization of the extragalactic reference system in optical and near infrared wavelengths

    Science.gov (United States)

    Akhmetov, Volodymyr S.; Fedorov, Peter N.; Velichko, Anna B.

    2018-04-01

    We combined the data from the Gaia DR1 and Two-Micron All Sky Survey (2MASS) catalogues in order to derive the absolute proper motions more than 420 million stars distributed all over the sky in the stellar magnitude range 8 mag 2MASS catalogue objects, the 2-dimensional median filter was used. The PMA system of proper motion has been obtained by direct link to 1.6 millions extragalactic sources. The short analysis of the absolute proper motion of the PMA stars Catalogue is presented in this work. From a comparison of this data with same stars from the TGAS, UCAC4 and PPMXL catalogues, the equatorial components of the mutual rotation vector of these coordinate systems are determined.

  4. Multi-source Geospatial Data Analysis with Google Earth Engine

    Science.gov (United States)

    Erickson, T.

    2014-12-01

    The Google Earth Engine platform is a cloud computing environment for data analysis that combines a public data catalog with a large-scale computational facility optimized for parallel processing of geospatial data. The data catalog is a multi-petabyte archive of georeferenced datasets that include images from Earth observing satellite and airborne sensors (examples: USGS Landsat, NASA MODIS, USDA NAIP), weather and climate datasets, and digital elevation models. Earth Engine supports both a just-in-time computation model that enables real-time preview and debugging during algorithm development for open-ended data exploration, and a batch computation mode for applying algorithms over large spatial and temporal extents. The platform automatically handles many traditionally-onerous data management tasks, such as data format conversion, reprojection, and resampling, which facilitates writing algorithms that combine data from multiple sensors and/or models. Although the primary use of Earth Engine, to date, has been the analysis of large Earth observing satellite datasets, the computational platform is generally applicable to a wide variety of use cases that require large-scale geospatial data analyses. This presentation will focus on how Earth Engine facilitates the analysis of geospatial data streams that originate from multiple separate sources (and often communities) and how it enables collaboration during algorithm development and data exploration. The talk will highlight current projects/analyses that are enabled by this functionality.https://earthengine.google.org

  5. The national atlas as a metaphor for improved use of a national geospatial data infrastructure

    NARCIS (Netherlands)

    Aditya Kurniawan Muhammad, T.

    2007-01-01

    Geospatial Data infrastructures have been developed worldwide. Geoportals have been created as an interface to allow users or the community to discover and use geospatial data offered by providers of these initiatives. This study focuses on the development of a web national atlas as an alternative

  6. Nebula observations. Catalogues and archive of photoplates

    Science.gov (United States)

    Shlyapnikov, A. A.; Smirnova, M. A.; Elizarova, N. V.

    2017-12-01

    A process of data systematization based on "Academician G.A. Shajn's Plan" for studying the Galaxy structure related to nebula observations is considered. The creation of digital versions of catalogues of observations and publications is described, as well as their presentation in HTML, VOTable and AJS formats and basic principles of work in the interactive application of International Virtual Observatory the Aladin Sky Atlas.

  7. Electromagnetic compatibility principles and applications

    CERN Document Server

    Weston, David A

    2001-01-01

    This totally revised and expanded reference/text provides comprehensive, single-source coverage of the design, problem solving, and specifications of electromagnetic compatibility (EMC) into electrical equipment/systems-including new information on basic theories, applications, evaluations, prediction techniques, and practical diagnostic options for preventing EMI through cost-effective solutions. Offers the most recent guidelines, safety limits, and standards for human exposure to electromagnetic fields! Containing updated data on EMI diagnostic verification measurements, as well as over 900 drawings, photographs, tables, and equations-500 more than the previous edition

  8. Radioisotopes and radiopharmaceuticals catalogue

    International Nuclear Information System (INIS)

    2002-01-01

    The Chilean Nuclear Energy Commission (CCHEN) presents its radioisotopes and radiopharmaceuticals 2002 catalogue. In it we found physical characteristics of 9 different reactor produced radioisotopes ( Tc-99m, I-131, Sm-153, Ir-192, P-32, Na-24, K-42, Cu-64, Rb-86 ), 7 radiopharmaceuticals ( MDP, DTPA, DMSA, Disida, Phitate, S-Coloid, Red Blood Cells In-Vivo, Red Blood Cells In-Vitro) and 4 labelled compounds ( DMSA-Tc99m, DTPA-Tc99m, MIBG-I131, EDTMP-Sm153 ). In the near future the number of items will be increased with new reactor and cyclotron products. Our production system will be certified by ISO 9000 on March 2003. CCHEN is interested in being a national and an international supplier of these products (RS)

  9. Technology catalogue. Second edition

    International Nuclear Information System (INIS)

    1995-04-01

    The Department of Energy's (DOE's) Office of Environmental Management (EM) is responsible for remediating DOE contaminated sites and managing the DOE waste inventory in a safe and efficient manner. EM's Office of Technology Development (OTD) supports applied research and demonstration efforts to develop and transfer innovative, cost-effective technologies to its site clean-up and waste-management programs within EM. The purpose of the Technology Catalogue is to: (a) provide performance data on OTD-developed technologies to scientists and engineers responsible for preparing Remedial Investigation/Feasibility Studies (RI/FSs) and other compliance documents for the DOE's clean-up and waste-management programs; and (b) identify partnering and commercialization opportunities with industry, other federal and state agencies, and the academic community

  10. Technology catalogue. Second edition

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1995-04-01

    The Department of Energy`s (DOE`s) Office of Environmental Management (EM) is responsible for remediating DOE contaminated sites and managing the DOE waste inventory in a safe and efficient manner. EM`s Office of Technology Development (OTD) supports applied research and demonstration efforts to develop and transfer innovative, cost-effective technologies to its site clean-up and waste-management programs within EM. The purpose of the Technology Catalogue is to: (a) provide performance data on OTD-developed technologies to scientists and engineers responsible for preparing Remedial Investigation/Feasibility Studies (RI/FSs) and other compliance documents for the DOE`s clean-up and waste-management programs; and (b) identify partnering and commercialization opportunities with industry, other federal and state agencies, and the academic community.

  11. High Performance Processing and Analysis of Geospatial Data Using CUDA on GPU

    Directory of Open Access Journals (Sweden)

    STOJANOVIC, N.

    2014-11-01

    Full Text Available In this paper, the high-performance processing of massive geospatial data on many-core GPU (Graphic Processing Unit is presented. We use CUDA (Compute Unified Device Architecture programming framework to implement parallel processing of common Geographic Information Systems (GIS algorithms, such as viewshed analysis and map-matching. Experimental evaluation indicates the improvement in performance with respect to CPU-based solutions and shows feasibility of using GPU and CUDA for parallel implementation of GIS algorithms over large-scale geospatial datasets.

  12. Global Compilation of InSAR Earthquake Source Models: Comparisons with Seismic Catalogues and the Effects of 3D Earth Structure

    Science.gov (United States)

    Weston, J. M.; Ferreira, A. M.; Funning, G. J.

    2010-12-01

    While past progress in seismology led to extensive earthquake catalogues such as the Global Centroid Moment Tensor (GCMT) catalogue, recent advances in space geodesy have enabled earthquake parameter estimations from the measurement of the deformation of the Earth’s surface, notably using InSAR data. Many earthquakes have now been studied using InSAR, but a full assessment of the quality and of the additional value of these source parameters compared to traditional seismological techniques is still lacking. In this study we present results of systematic comparisons between earthquake CMT parameters determined using InSAR and seismic data, on a global scale. We compiled a large database of source parameters obtained using InSAR data from the literature and estimated the corresponding CMT parameters into a ICMT compilation. We here present results from the analysis of 58 earthquakes that occurred between 1992-2007 from about 80 published InSAR studies. Multiple studies of the same earthquake are included in the archive, as they are valuable to assess uncertainties. Where faults are segmented, with changes in width along-strike, a weighted average based on the seismic moment in each fault has been used to determine overall earthquake parameters. For variable slip models, we have calculated source parameters taking the spatial distribution of slip into account. The parameters in our ICMT compilation are compared with those taken from the Global CMT (GCMT), ISC, EHB and NEIC catalogues. We find that earthquake fault strike, dip and rake values in the GCMT and ICMT archives are generally compatible with each other. Likewise, the differences in seismic moment in these two archives are relatively small. However, the locations of the centroid epicentres show substantial discrepancies, which are larger when comparing with GCMT locations (10-30km differences) than for EHB, ISC and NEIC locations (5-15km differences). Since InSAR data have a high spatial resolution, and thus

  13. Catalogue of nuclear fusion codes - 1976

    International Nuclear Information System (INIS)

    1976-10-01

    A catalogue is presented of the computer codes in nuclear fusion research developed by JAERI, Division of Thermonuclear Fusion Research and Division of Large Tokamak Development in particular. It contains a total of about 100 codes under the categories: Atomic Process, Data Handling, Experimental Data Processing, Engineering, Input and Output, Special Languages and Their Application, Mathematical Programming, Miscellaneous, Numerical Analysis, Nuclear Physics, Plasma Physics and Fusion Research, Plasma Simulation and Numerical Technique, Reactor Design, Solid State Physics, Statistics, and System Program. (auth.)

  14. fMRI-compatible rehabilitation hand device

    Directory of Open Access Journals (Sweden)

    Tzika Aria

    2006-10-01

    Full Text Available Abstract Background Functional magnetic resonance imaging (fMRI has been widely used in studying human brain functions and neurorehabilitation. In order to develop complex and well-controlled fMRI paradigms, interfaces that can precisely control and measure output force and kinematics of the movements in human subjects are needed. Optimized state-of-the-art fMRI methods, combined with magnetic resonance (MR compatible robotic devices for rehabilitation, can assist therapists to quantify, monitor, and improve physical rehabilitation. To achieve this goal, robotic or mechatronic devices with actuators and sensors need to be introduced into an MR environment. The common standard mechanical parts can not be used in MR environment and MR compatibility has been a tough hurdle for device developers. Methods This paper presents the design, fabrication and preliminary testing of a novel, one degree of freedom, MR compatible, computer controlled, variable resistance hand device that may be used in brain MR imaging during hand grip rehabilitation. We named the device MR_CHIROD (Magnetic Resonance Compatible Smart Hand Interfaced Rehabilitation Device. A novel feature of the device is the use of Electro-Rheological Fluids (ERFs to achieve tunable and controllable resistive force generation. ERFs are fluids that experience dramatic changes in rheological properties, such as viscosity or yield stress, in the presence of an electric field. The device consists of four major subsystems: a an ERF based resistive element; b a gearbox; c two handles and d two sensors, one optical encoder and one force sensor, to measure the patient induced motion and force. The smart hand device is designed to resist up to 50% of the maximum level of gripping force of a human hand and be controlled in real time. Results Laboratory tests of the device indicate that it was able to meet its design objective to resist up to approximately 50% of the maximum handgrip force. The detailed

  15. Mapping Heritage: Geospatial Online Databases of Historic Roads. The Case of the N-340 Roadway Corridor on the Spanish Mediterranean

    Directory of Open Access Journals (Sweden)

    Mar Loren-Méndez

    2018-04-01

    Full Text Available The study has developed an online geospatial database for assessing the complexity of roadway heritage, overcoming the limitations of traditional heritage catalogues and databases: the itemization of heritage assets and the rigidity of the database structure. Reflecting the current openness in the field of heritage studies, the research proposes an interdisciplinary approach that reframes heritage databases, both conceptually and technologically. Territorial scale is key for heritage interpretation, the complex characteristics of each type of heritage, and social appropriation. The system is based on an open-source content-management system and framework called ProcessWire, allowing flexibility in the definition of data fields and serving as an internal working tool for research collaboration. Accessibility, flexibility, and ease of use do not preclude rigor: the database works in conjunction with a GIS (Geographic Information System support system and is complemented by a bibliographical archive. A hierarchical multiscalar heritage characterization has been implemented in order to include the different territorial scales and to facilitate the creation of itineraries. Having attained the main goals of conceptual heritage coherence, accessibility, and rigor, the database should strive for broader capacity to integrate GIS information and stimulate public participation, a step toward controlled crowdsourcing and collaborative heritage characterization.

  16. Automated Geospatial Watershed Assessment Tool (AGWA) Poster Presentation

    Science.gov (United States)

    The Automated Geospatial Watershed Assessment tool (AGWA, see: www.tucson.ars.ag.gov/agwa or http://www.epa.gov/esd/land-sci/agwa/) is a GIS interface jointly developed by the USDA-Agricultural Research Service, the U.S. Environmental Protection Agency, the University of Arizona...

  17. A Research Agenda for Geospatial Technologies and Learning

    Science.gov (United States)

    Baker, Tom R.; Battersby, Sarah; Bednarz, Sarah W.; Bodzin, Alec M.; Kolvoord, Bob; Moore, Steven; Sinton, Diana; Uttal, David

    2015-01-01

    Knowledge around geospatial technologies and learning remains sparse, inconsistent, and overly anecdotal. Studies are needed that are better structured; more systematic and replicable; attentive to progress and findings in the cognate fields of science, technology, engineering, and math education; and coordinated for multidisciplinary approaches.…

  18. Academic research opportunities at the National Geospatial-Intelligence Agency(NGA)

    Science.gov (United States)

    Loomer, Scott A.

    2006-05-01

    The vision of the National Geospatial-Intelligence Agency (NGA) is to "Know the Earth...Show the Way." To achieve this vision, the NGA provides geospatial intelligence in all its forms and from whatever source-imagery, imagery intelligence, and geospatial data and information-to ensure the knowledge foundation for planning, decision, and action. Academia plays a key role in the NGA research and development program through the NGA Academic Research Program. This multi-disciplinary program of basic research in geospatial intelligence topics provides grants and fellowships to the leading investigators, research universities, and colleges of the nation. This research provides the fundamental science support to NGA's applied and advanced research programs. The major components of the NGA Academic Research Program are: *NGA University Research Initiatives (NURI): Three-year basic research grants awarded competitively to the best investigators across the US academic community. Topics are selected to provide the scientific basis for advanced and applied research in NGA core disciplines. *Historically Black College and University - Minority Institution Research Initiatives (HBCU-MI): Two-year basic research grants awarded competitively to the best investigators at Historically Black Colleges and Universities, and Minority Institutions across the US academic community. *Intelligence Community Post-Doctoral Research Fellowships: Fellowships providing access to advanced research in science and technology applicable to the intelligence community's mission. The program provides a pool of researchers to support future intelligence community needs and develops long-term relationships with researchers as they move into career positions. This paper provides information about the NGA Academic Research Program, the projects it supports and how researchers and institutions can apply for grants under the program. In addition, other opportunities for academia to engage with NGA through

  19. Automation of electromagnetic compatability (EMC) test facilities

    Science.gov (United States)

    Harrison, C. A.

    1986-01-01

    Efforts to automate electromagnetic compatibility (EMC) test facilities at Marshall Space Flight Center are discussed. The present facility is used to accomplish a battery of nine standard tests (with limited variations) deigned to certify EMC of Shuttle payload equipment. Prior to this project, some EMC tests were partially automated, but others were performed manually. Software was developed to integrate all testing by means of a desk-top computer-controller. Near real-time data reduction and onboard graphics capabilities permit immediate assessment of test results. Provisions for disk storage of test data permit computer production of the test engineer's certification report. Software flexibility permits variation in the tests procedure, the ability to examine more closely those frequency bands which indicate compatibility problems, and the capability to incorporate additional test procedures.

  20. Conceptual models in the field of library catalogues

    Directory of Open Access Journals (Sweden)

    Marija Petek

    2000-01-01

    Full Text Available The publishing world is changing quickly and so must also bibliographic control. It is tirne to re-examine cataloguing rules and MARC formats. This can be done by the method of conceptual modelling. Some conceptual models are presented; an IFLA study on the functional requirements for bibliographic records is described in detail.

  1. Improving Library Management by Using Cost Analysis Tools: A Case Study for Cataloguing Processes

    Directory of Open Access Journals (Sweden)

    Lorena Siguenza-Guzman

    2014-02-01

    Full Text Available TTDABC is a relatively new costing management technique, initially developed for manufacturing processes, which is gaining attention in libraries. This is because TDABC is a fast and simple method that only requires two parameters, an estimation of time required to perform an activity and the unit cost per time of supplying capacity. A few case studies have been documented with regard to TDABC in libraries; all of them being oriented to analyse specific library activities such as inter-library loan, acquisition and circulation processes. The primary focus of this paper is to describe TDABC implementation in one of the most important library processes, namely cataloguing. In particular, original and copy cataloguing are analysed through a case study to demonstrate the applicability and usefulness of TDABC to perform cost analysis of cataloguing processes.

  2. Geospatial cryptography: enabling researchers to access private, spatially referenced, human subjects data for cancer control and prevention.

    Science.gov (United States)

    Jacquez, Geoffrey M; Essex, Aleksander; Curtis, Andrew; Kohler, Betsy; Sherman, Recinda; Emam, Khaled El; Shi, Chen; Kaufmann, Andy; Beale, Linda; Cusick, Thomas; Goldberg, Daniel; Goovaerts, Pierre

    2017-07-01

    As the volume, accuracy and precision of digital geographic information have increased, concerns regarding individual privacy and confidentiality have come to the forefront. Not only do these challenge a basic tenet underlying the advancement of science by posing substantial obstacles to the sharing of data to validate research results, but they are obstacles to conducting certain research projects in the first place. Geospatial cryptography involves the specification, design, implementation and application of cryptographic techniques to address privacy, confidentiality and security concerns for geographically referenced data. This article defines geospatial cryptography and demonstrates its application in cancer control and surveillance. Four use cases are considered: (1) national-level de-duplication among state or province-based cancer registries; (2) sharing of confidential data across cancer registries to support case aggregation across administrative geographies; (3) secure data linkage; and (4) cancer cluster investigation and surveillance. A secure multi-party system for geospatial cryptography is developed. Solutions under geospatial cryptography are presented and computation time is calculated. As services provided by cancer registries to the research community, de-duplication, case aggregation across administrative geographies and secure data linkage are often time-consuming and in some instances precluded by confidentiality and security concerns. Geospatial cryptography provides secure solutions that hold significant promise for addressing these concerns and for accelerating the pace of research with human subjects data residing in our nation's cancer registries. Pursuit of the research directions posed herein conceivably would lead to a geospatially encrypted geographic information system (GEGIS) designed specifically to promote the sharing and spatial analysis of confidential data. Geospatial cryptography holds substantial promise for accelerating the

  3. Big Data analytics in the Geo-Spatial Domain

    NARCIS (Netherlands)

    R.A. Goncalves (Romulo); M.G. Ivanova (Milena); M.L. Kersten (Martin); H. Scholten; S. Zlatanova; F. Alvanaki (Foteini); P. Nourian (Pirouz); E. Dias

    2014-01-01

    htmlabstractBig data collections in many scientific domains have inherently rich spatial and geo-spatial features. Spatial location is among the core aspects of data in Earth observation sciences, astronomy, and seismology to name a few. The goal of our project is to design an efficient data

  4. Intelligence, mapping, and geospatial exploitation system (IMAGES)

    Science.gov (United States)

    Moellman, Dennis E.; Cain, Joel M.

    1998-08-01

    This paper provides further detail to one facet of the battlespace visualization concept described in last year's paper Battlespace Situation Awareness for Force XXI. It focuses on the National Imagery and Mapping Agency (NIMA) goal to 'provide customers seamless access to tailorable imagery, imagery intelligence, and geospatial information.' This paper describes Intelligence, Mapping, and Geospatial Exploitation System (IMAGES), an exploitation element capable of CONUS baseplant operations or field deployment to provide NIMA geospatial information collaboratively into a reconnaissance, surveillance, and target acquisition (RSTA) environment through the United States Imagery and Geospatial Information System (USIGS). In a baseplant CONUS setting IMAGES could be used to produce foundation data to support mission planning. In the field it could be directly associated with a tactical sensor receiver or ground station (e.g. UAV or UGV) to provide near real-time and mission specific RSTA to support mission execution. This paper provides IMAGES functional level design; describes the technologies, their interactions and interdependencies; and presents a notional operational scenario to illustrate the system flexibility. Using as a system backbone an intelligent software agent technology, called Open Agent ArchitectureTM (OAATM), IMAGES combines multimodal data entry, natural language understanding, and perceptual and evidential reasoning for system management. Configured to be DII COE compliant, it would utilize, to the extent possible, COTS applications software for data management, processing, fusion, exploitation, and reporting. It would also be modular, scaleable, and reconfigurable. This paper describes how the OAATM achieves data synchronization and enables the necessary level of information to be rapidly available to various command echelons for making informed decisions. The reasoning component will provide for the best information to be developed in the timeline

  5. Cartographic standards to improve maps produced by the Forest Inventory and Analysis program

    Science.gov (United States)

    Charles H. (Hobie) Perry; Mark D. Nelson

    2009-01-01

    The Forest Service, U.S. Department of Agriculture's Forest Inventory and Analysis (FIA) program is incorporating an increasing number of cartographic products in reports, publications, and presentations. To create greater quality and consistency within the national FIA program, a Geospatial Standards team developed cartographic design standards for FIA map...

  6. Strengthened IAEA Safeguards-Imagery Analysis: Geospatial Tools for Nonproliferation Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Pabian, Frank V [Los Alamos National Laboratory

    2012-08-14

    This slide presentation focuses on the growing role and importance of imagery analysis for IAEA safeguards applications and how commercial satellite imagery, together with the newly available geospatial tools, can be used to promote 'all-source synergy.' As additional sources of openly available information, satellite imagery in conjunction with the geospatial tools can be used to significantly augment and enhance existing information gathering techniques, procedures, and analyses in the remote detection and assessment of nonproliferation relevant activities, facilities, and programs. Foremost of the geospatial tools are the 'Digital Virtual Globes' (i.e., GoogleEarth, Virtual Earth, etc.) that are far better than previously used simple 2-D plan-view line drawings for visualization of known and suspected facilities of interest which can be critical to: (1) Site familiarization and true geospatial context awareness; (2) Pre-inspection planning; (3) Onsite orientation and navigation; (4) Post-inspection reporting; (5) Site monitoring over time for changes; (6) Verification of states site declarations and for input to State Evaluation reports; and (7) A common basis for discussions among all interested parties (Member States). Additionally, as an 'open-source', such virtual globes can also provide a new, essentially free, means to conduct broad area search for undeclared nuclear sites and activities - either alleged through open source leads; identified on internet BLOGS and WIKI Layers, with input from a 'free' cadre of global browsers and/or by knowledgeable local citizens (a.k.a.: 'crowdsourcing'), that can include ground photos and maps; or by other initiatives based on existing information and in-house country knowledge. They also provide a means to acquire ground photography taken by locals, hobbyists, and tourists of the surrounding locales that can be useful in identifying and discriminating between relevant

  7. IMPRINT Analysis of an Unmanned Air System Geospatial Information Process

    National Research Council Canada - National Science Library

    Hunn, Bruce P; Schweitzer, Kristin M; Cahir, John A; Finch, Mary M

    2008-01-01

    ... intelligence, geospatial analysis cell. The Improved Performance Research Integration Tool (IMPRINT) modeling program was used to understand this process and to assess crew workload during several test scenarios...

  8. Hipparcos to deliver its final results catalogue soon

    Science.gov (United States)

    1995-10-01

    them, almost 30 years ago, to propose carrying out these observations from the relatively benign environment of space. Hipparcos is, by present standards, a medium-sized satellite, with a 30 cm telescope sensing simply ordinary light. But it has been described as the most imaginative in the short history of space astronomy. This foresight has been amply repaid. In the long history of stargazing it ranks with the surveys by Hipparchus the Greek in the 2nd Century BC and by Tichy Brahe the Dane in the 16th Century AD, both of which transformed human perceptions of the Universe. Positions derived from the Hipparcos satellite are better than a millionth of a degree, and newly a thousand times more accurate than star positions routinely determined from he ground. This accuracy makes it possible to measure directly the distances to the stars. While it took 250 years between astronomers first setting out on the exacting task of measuring the distance to a star, and a stellar distance being measured for the first time, ESA's Hipparcos mission has revolutionised this long, painstaking, and fundamental task by measuring accurate distances and movements of more than one hundred thousand. The measurement concept involved he satellite triangulating its way between he stars all wound the sky, building up a celestial map in much the same way as land surveyors use triangulation between hill-tops to measure distances accurately. Only the angles involved are much smaller : the accuracy that has been achieved with the Hipparcos Catalogue is such that he two edges of a coin, viewed from he other side of the Atlantic Ocean, could be distinguished. The results from Hipparcos will deliver scientists with long-awaited details of our place in he Milky Way Galaxy. Most of he stars visible to the naked eye are, to a large extent, companions of the Sun, in a great orbital march around the centre of the Galaxy, a journey so long that it takes individual stars 250 million years to complete, in

  9. Learning R for geospatial analysis

    CERN Document Server

    Dorman, Michael

    2014-01-01

    This book is intended for anyone who wants to learn how to efficiently analyze geospatial data with R, including GIS analysts, researchers, educators, and students who work with spatial data and who are interested in expanding their capabilities through programming. The book assumes familiarity with the basic geographic information concepts (such as spatial coordinates), but no prior experience with R and/or programming is required. By focusing on R exclusively, you will not need to depend on any external software-a working installation of R is all that is necessary to begin.

  10. [Criteria catalogue to systematize conceptual approaches in universal prevention of childhood overweight : Methodological approach and first results].

    Science.gov (United States)

    Babitsch, Birgit; Geene, Raimund; Hassel, Holger; Kliche, Thomas; Bacchetta, Britta; Baltes, Simon; Nold, Sandra; Rosenfeldt, Daniel

    2016-11-01

    Overweight and obesity are serious health risks for children and adolescents. Hence, various prevention projects have been initiated and implemented. Until now, a systematic overview of interventions in different settings has been lacking. The aim of the "Prevention of child overweight" project (SkAP-project) is to prepare a systematic overview of the conceptual approaches used in universal prevention of overweight among children and adolescents. First of all, a comprehensive criteria catalogue will be developed based on systematic searches. In the next step the criteria catalogue will be applied to identify and characterize conceptual approaches. Criteria to describe conceptual approaches as well as determinants of childhood overweight were determined by systematic searches. The searches included relevant data bases and were further expanded by internet and hand search. Three settings (kindergarten, school and communities) and families are addressed by the systematic searches. Additional non-setting specific searches were conducted. A comprehensive criteria catalogue was developed, which allows a detailed analysis of conceptual approaches. This catalogue covers further quality criteria as well as determinants of childhood overweight. Currently, the criteria catalogue is being employed. Although the detailed analysis of conceptual approaches can be regarded as advantage of the criteria catalogue, there are also some limitations, such as the lack of necessary information provided in publications. Overall, the application will reveal an overview regarding universal prevention in childhood overweight, which is still lacking, and will support development in this field.

  11. Electronic spare-parts catalogue, an elementary module of spare-parts supply; Elektronischer Ersatzteilkatalog als elementarer Baustein der Ersatzteilversorgung

    Energy Technology Data Exchange (ETDEWEB)

    Nienhaus, Karl; Bartnitzki, Thomas [RWTH Aachen (Germany). Inst. fuer Maschinentechnik der Rohstoffindustrie; Stoll, Andrea [Bucyrus HEX GmbH, Dortmund (Germany). Abt. Technische Dokumentation

    2011-01-15

    Developments in recent years in the supply of spare parts at Bucyrus HEX GmbH have revealed the increasing importance of quick and error-free ordering of spare parts by means of electronic spare-parts catalogues. The spare-parts catalogue has meanwhile become the central component in order to enter eCommerce with the after sales services. These higher requirements will lead to further optimisation and automation of the publishing process on the manufacturer's side and to quicker acceptance of the new digital media and the linking of the electronic spare-parts catalogues to maintenance planning systems on the customer's side. The authors agree that these are exciting times with regard to electronic spare-parts catalogues. (orig.)

  12. A Geospatial Cyberinfrastructure for Urban Economic Analysis and Spatial Decision-Making

    Directory of Open Access Journals (Sweden)

    Michael F. Goodchild

    2013-05-01

    Full Text Available Urban economic modeling and effective spatial planning are critical tools towards achieving urban sustainability. However, in practice, many technical obstacles, such as information islands, poor documentation of data and lack of software platforms to facilitate virtual collaboration, are challenging the effectiveness of decision-making processes. In this paper, we report on our efforts to design and develop a geospatial cyberinfrastructure (GCI for urban economic analysis and simulation. This GCI provides an operational graphic user interface, built upon a service-oriented architecture to allow (1 widespread sharing and seamless integration of distributed geospatial data; (2 an effective way to address the uncertainty and positional errors encountered in fusing data from diverse sources; (3 the decomposition of complex planning questions into atomic spatial analysis tasks and the generation of a web service chain to tackle such complex problems; and (4 capturing and representing provenance of geospatial data to trace its flow in the modeling task. The Greater Los Angeles Region serves as the test bed. We expect this work to contribute to effective spatial policy analysis and decision-making through the adoption of advanced GCI and to broaden the application coverage of GCI to include urban economic simulations.

  13. Bridging the Gap between NASA Hydrological Data and the Geospatial Community

    Science.gov (United States)

    Rui, Hualan; Teng, Bill; Vollmer, Bruce; Mocko, David M.; Beaudoing, Hiroko K.; Nigro, Joseph; Gary, Mark; Maidment, David; Hooper, Richard

    2011-01-01

    There is a vast and ever increasing amount of data on the Earth interconnected energy and hydrological systems, available from NASA remote sensing and modeling systems, and yet, one challenge persists: increasing the usefulness of these data for, and thus their use by, the geospatial communities. The Hydrology Data and Information Services Center (HDISC), part of the Goddard Earth Sciences DISC, has continually worked to better understand the hydrological data needs of the geospatial end users, to thus better able to bridge the gap between NASA data and the geospatial communities. This paper will cover some of the hydrological data sets available from HDISC, and the various tools and services developed for data searching, data subletting ; format conversion. online visualization and analysis; interoperable access; etc.; to facilitate the integration of NASA hydrological data by end users. The NASA Goddard data analysis and visualization system, Giovanni, is described. Two case examples of user-customized data services are given, involving the EPA BASINS (Better Assessment Science Integrating point & Non-point Sources) project and the CUAHSI Hydrologic Information System, with the common requirement of on-the-fly retrieval of long duration time series for a geographical point

  14. Retrospective Conversion of Card catalogue at the University of ...

    African Journals Online (AJOL)

    Also, the article describes the conversion methods employed, procedures followed, and pre-conversion preparations made by the library to automate its card catalogue and to assign barcode labels to the collection. Finally it concludes by highlighting problems encountered throughout the project and by giving statistical ...

  15. Energy renovation solutions - catalogue; Energirenoveringstiltag - katalog

    Energy Technology Data Exchange (ETDEWEB)

    Tommerup, H.

    2010-07-15

    The project's aim has been to develop methods and examples of extensive energy renovations to stimulate energy conservation and increased use of renewable energy in existing buildings. The current report represents an extensive technology catalogue of typical energy renovation measures in connection with the renovation of existing buildings. For every action the main aspects are explained concerning such issues as technology, use, barriers, indoor climate, energy conservation and prices. The report is mainly targeted at construction industry, but also many other stakeholders can benefit from the report. (ln)

  16. Creating 3D models of historical buildings using geospatial data

    Science.gov (United States)

    Alionescu, Adrian; Bǎlǎ, Alina Corina; Brebu, Floarea Maria; Moscovici, Anca-Maria

    2017-07-01

    Recently, a lot of interest has been shown to understand a real world object by acquiring its 3D images of using laser scanning technology and panoramic images. A realistic impression of geometric 3D data can be generated by draping real colour textures simultaneously captured by a colour camera images. In this context, a new concept of geospatial data acquisition has rapidly revolutionized the method of determining the spatial position of objects, which is based on panoramic images. This article describes an approach that comprises inusing terrestrial laser scanning and panoramic images captured with Trimble V10 Imaging Rover technology to enlarge the details and realism of the geospatial data set, in order to obtain 3D urban plans and virtual reality applications.

  17. Persistent Teaching Practices after Geospatial Technology Professional Development

    Science.gov (United States)

    Rubino-Hare, Lori A.; Whitworth, Brooke A.; Bloom, Nena E.; Claesgens, Jennifer M.; Fredrickson, Kristi M.; Sample, James C.

    2016-01-01

    This case study described teachers with varying technology skills who were implementing the use of geospatial technology (GST) within project-based instruction (PBI) at varying grade levels and contexts 1 to 2 years following professional development. The sample consisted of 10 fifth- to ninth-grade teachers. Data sources included artifacts,…

  18. A Catalogue of the General Electric Photographic Archives, 1890-1940

    DEFF Research Database (Denmark)

    Nye, David Edwin

    Catalogue of 5,000 selected images from the GE photographic Archives, together with an introduction and Index. c. 200 photographs reproduced in the book. A revised edition was published in 1998, but without the original author's involvement....

  19. Lsiviewer 2.0 - a Client-Oriented Online Visualization Tool for Geospatial Vector Data

    Science.gov (United States)

    Manikanta, K.; Rajan, K. S.

    2017-09-01

    Geospatial data visualization systems have been predominantly through applications that are installed and run in a desktop environment. Over the last decade, with the advent of web technologies and its adoption by Geospatial community, the server-client model for data handling, data rendering and visualization respectively has been the most prevalent approach in Web-GIS. While the client devices have become functionally more powerful over the recent years, the above model has largely ignored it and is still in a mode of serverdominant computing paradigm. In this paper, an attempt has been made to develop and demonstrate LSIViewer - a simple, easy-to-use and robust online geospatial data visualisation system for the user's own data that harness the client's capabilities for data rendering and user-interactive styling, with a reduced load on the server. The developed system can support multiple geospatial vector formats and can be integrated with other web-based systems like WMS, WFS, etc. The technology stack used to build this system is Node.js on the server side and HTML5 Canvas and JavaScript on the client side. Various tests run on a range of vector datasets, upto 35 MB, showed that the time taken to render the vector data using LSIViewer is comparable to a desktop GIS application, QGIS, over an identical system.

  20. A Geo-Event-Based Geospatial Information Service: A Case Study of Typhoon Hazard

    Directory of Open Access Journals (Sweden)

    Yu Zhang

    2017-03-01

    Full Text Available Social media is valuable in propagating information during disasters for its timely and available characteristics nowadays, and assists in making decisions when tagged with locations. Considering the ambiguity and inaccuracy in some social data, additional authoritative data are needed for important verification. However, current works often fail to leverage both social and authoritative data and, on most occasions, the data are used in disaster analysis after the fact. Moreover, current works organize the data from the perspective of the spatial location, but not from the perspective of the disaster, making it difficult to dynamically analyze the disaster. All of the disaster-related data around the affected locations need to be retrieved. To solve these limitations, this study develops a geo-event-based geospatial information service (GEGIS framework and proceeded as follows: (1 a geo-event-related ontology was constructed to provide a uniform semantic basis for the system; (2 geo-events and attributes were extracted from the web using a natural language process (NLP and used in the semantic similarity match of the geospatial resources; and (3 a geospatial information service prototype system was designed and implemented for automatically retrieving and organizing geo-event-related geospatial resources. A case study of a typhoon hazard is analyzed here within the GEGIS and shows that the system would be effective when typhoons occur.

  1. Geospatial Technology: A Tool to Aid in the Elimination of Malaria in Bangladesh

    Directory of Open Access Journals (Sweden)

    Karen E. Kirk

    2014-12-01

    Full Text Available Bangladesh is a malaria endemic country. There are 13 districts in the country bordering India and Myanmar that are at risk of malaria. The majority of malaria morbidity and mortality cases are in the Chittagong Hill Tracts, the mountainous southeastern region of Bangladesh. In recent years, malaria burden has declined in the country. In this study, we reviewed and summarized published data (through 2014 on the use of geospatial technologies on malaria epidemiology in Bangladesh and outlined potential contributions of geospatial technologies for eliminating malaria in the country. We completed a literature review using “malaria, Bangladesh” search terms and found 218 articles published in peer-reviewed journals listed in PubMed. After a detailed review, 201 articles were excluded because they did not meet our inclusion criteria, 17 articles were selected for final evaluation. Published studies indicated geospatial technologies tools (Geographic Information System, Global Positioning System, and Remote Sensing were used to determine vector-breeding sites, land cover classification, accessibility to health facility, treatment seeking behaviors, and risk mapping at the household, regional, and national levels in Bangladesh. To achieve the goal of malaria elimination in Bangladesh, we concluded that further research using geospatial technologies should be integrated into the country’s ongoing surveillance system to identify and better assess progress towards malaria elimination.

  2. Designing a two-rank acceptance sampling plan for quality inspection of geospatial data products

    Science.gov (United States)

    Tong, Xiaohua; Wang, Zhenhua; Xie, Huan; Liang, Dan; Jiang, Zuoqin; Li, Jinchao; Li, Jun

    2011-10-01

    To address the disadvantages of classical sampling plans designed for traditional industrial products, we originally propose a two-rank acceptance sampling plan (TRASP) for the inspection of geospatial data outputs based on the acceptance quality level (AQL). The first rank sampling plan is to inspect the lot consisting of map sheets, and the second is to inspect the lot consisting of features in an individual map sheet. The TRASP design is formulated as an optimization problem with respect to sample size and acceptance number, which covers two lot size cases. The first case is for a small lot size with nonconformities being modeled by a hypergeometric distribution function, and the second is for a larger lot size with nonconformities being modeled by a Poisson distribution function. The proposed TRASP is illustrated through two empirical case studies. Our analysis demonstrates that: (1) the proposed TRASP provides a general approach for quality inspection of geospatial data outputs consisting of non-uniform items and (2) the proposed acceptance sampling plan based on TRASP performs better than other classical sampling plans. It overcomes the drawbacks of percent sampling, i.e., "strictness for large lot size, toleration for small lot size," and those of a national standard used specifically for industrial outputs, i.e., "lots with different sizes corresponding to the same sampling plan."

  3. Theoretical multi-tier trust framework for the geospatial domain

    CSIR Research Space (South Africa)

    Umuhoza, D

    2010-01-01

    Full Text Available chain or workflow from data acquisition to knowledge discovery. The author’s present work in progress of a theoretical multi-tier trust framework for processing chain from data acquisition to knowledge discovery in geospatial domain. Holistic trust...

  4. Geospatial big data and cartography : research challenges and opportunities for making maps that matter

    OpenAIRE

    Robinson, Anthony C.; Demsar, Urska; Moore, Antoni B.; Buckley, Aileen; Jiang, Bin; Field, Kenneth; Kraak, Menno-Jan; Camboim, Silvana P; Sluter, Claudia R

    2017-01-01

    Geospatial big data present a new set of challenges and opportunities for cartographic researchers in technical, methodological, and artistic realms. New computational and technical paradigms for cartography are accompanying the rise of geospatial big data. Additionally, the art and science of cartography needs to focus its contemporary efforts on work that connects to outside disciplines and is grounded in problems that are important to humankind and its sustainability. Following the develop...

  5. Modelling and Implementation of Catalogue Cards Using FreeMarker

    Science.gov (United States)

    Radjenovic, Jelen; Milosavljevic, Branko; Surla, Dusan

    2009-01-01

    Purpose: The purpose of this paper is to report on a study involving the specification (using Unified Modelling Language (UML) 2.0) of information requirements and implementation of the software components for generating catalogue cards. The implementation in a Java environment is developed using the FreeMarker software.…

  6. International Catalogue of Sealed Radioactive Sources and Devices

    International Nuclear Information System (INIS)

    2010-01-01

    The international catalogue of sealed radioactive sources and devices have two major objectives. The first objective is to provide vital information for a wide range of individuals and organizations on industrially manufactured radioactive sources and devices. The second objective is to facilitate identification of design specifications based on limited information from orphan sources and devices to allow safe handling of these items.

  7. A research on the security of wisdom campus based on geospatial big data

    Science.gov (United States)

    Wang, Haiying

    2018-05-01

    There are some difficulties in wisdom campus, such as geospatial big data sharing, function expansion, data management, analysis and mining geospatial big data for a characteristic, especially the problem of data security can't guarantee cause prominent attention increasingly. In this article we put forward a data-oriented software architecture which is designed by the ideology of orienting data and data as kernel, solve the problem of traditional software architecture broaden the campus space data research, develop the application of wisdom campus.

  8. A research project to encourage system-compatible design of end-use appliances

    International Nuclear Information System (INIS)

    Dorr, D.; Key, T.; Sitzlar, G.

    1995-01-01

    Cooperative system compatibility research sponsored by the Canadian Electrical Association (CEA) and the Electric Power Research Institute (EPRI) for improving appliance performance deficiencies was described. Power producer and end-user compatibility concerns was addressed through the development of a System Compatibility Research Project. A list of project tasks was provided. The CEA and EPRI initiated a project to establish flicker response of various lighting systems, which included physical tests. Results of this project were presented and discussed. The incentives for developing switch mode power supplies with enhanced immunity to voltage fluctuations and short interruptions was discussed. It was concluded that power quality studies currently underway will provide designers with a profile of the expected utility environment for their products. System compatibility research will identify areas that should be addressed by standards bodies so that designers can apply applicable criteria objectives early in the appliance design process. These efforts were expected to encourage appropriate manufacturer criteria for compatibility by convincing buyers and sellers that there is a real pay back for this investment. 13 refs., 6 figs., 4 tabs

  9. Research on presentation and query service of geo-spatial data based on ontology

    Science.gov (United States)

    Li, Hong-wei; Li, Qin-chao; Cai, Chang

    2008-10-01

    The paper analyzed the deficiency on presentation and query of geo-spatial data existed in current GIS, discussed the advantages that ontology possessed in formalization of geo-spatial data and the presentation of semantic granularity, taken land-use classification system as an example to construct domain ontology, and described it by OWL; realized the grade level and category presentation of land-use data benefited from the thoughts of vertical and horizontal navigation; and then discussed query mode of geo-spatial data based on ontology, including data query based on types and grade levels, instances and spatial relation, and synthetic query based on types and instances; these methods enriched query mode of current GIS, and is a useful attempt; point out that the key point of the presentation and query of spatial data based on ontology is to construct domain ontology that can correctly reflect geo-concept and its spatial relation and realize its fine formalization description.

  10. Are Higher Standards and Students' Needs Compatible?

    Science.gov (United States)

    Dodd, Anne Wescott

    2000-01-01

    Abraham Maslow's needs hierarchy proposes that people must satisfy their "deficiency" needs (for food, safety, love, belonging, and esteem) before they can move toward growth and achievement. Schools should help all students meet standards, stress rate of improvement, and report scores on school-climate measures. (MLH)

  11. VizieR Online Data Catalog: XMM-Newton slew survey Source Catalogue, version 2.0 (XMM-SSC, 2017)

    Science.gov (United States)

    XMM-SSC

    2018-01-01

    XMMSL2 is the second catalogue of X-ray sources found in slew data taken from the European Space Agency's (ESA) XMM-Newton observatory, and has been constructed by members of the XMM SOC and the EPIC consortium on behalf of ESA. This release uses results of work which has been carried out within the framework of the EXTraS project ("Exploring the X-ray variable and Transient Sky"), funded from the EU's Seventh Framework Programme under grant agreement no.607452. This is the first release of XMMSL2 which contains data taken between revolutions 314 and 2758. The previous catalogue was called XMMSL1_Delta6 and contained slews up to revolution 2441. The release includes two FITS files. A full catalogue (xmmsl2_total.fits.gz), containing 72352 detections found with a likelihood of DETML>8 and a "clean" catalogue (xmmsl2clean.fits.gz) where all known bad sources have been removed and where the detection limit has been raised to DETML>10.5 in general and DETML>15.5 for sources found in images with a higher than usual background. Efforts have been made to identify spurious detections and 3017 have been flagged as such in the full catalogue. (3 data files).

  12. National Recordal System IK holder catalogue process

    CSIR Research Space (South Africa)

    Pretorius, R

    2012-10-01

    Full Text Available good, based on IK.? Subsequently, the South African Department of Science and Technology (DST) created a National Indigenous Knowledge Systems Office (NIKSO) that is taking the lead on interfacing IKS with other recognised knowledge holders... is to illustrate the South African National Recordal Systems (NRS) process whereby Indigenous Knowledge (IK) holder?s information is captured in a registry, as specified in South Africa?s Indigenous Knowledge Systems (IKS) Policy, through an IK cataloguing...

  13. Short notes and reviews Birds in the sales catalogue of Adriaan Vroeg described by Pallas and Vosmaer

    NARCIS (Netherlands)

    Rookmaaker, L.C.; Pieters, Florence F.J.M.

    2000-01-01

    On 6 October 1764, Adriaan Vroeg sold his collection of birds, insects and a few other animals by auction. The sales catalogue was published anonymously, but contained several scientific names of birds, which have nomenclatorial standing. Only three copies of this catalogue are known. The appendix

  14. Improving the Slum Planning Through Geospatial Decision Support System

    Science.gov (United States)

    Shekhar, S.

    2014-11-01

    In India, a number of schemes and programmes have been launched from time to time in order to promote integrated city development and to enable the slum dwellers to gain access to the basic services. Despite the use of geospatial technologies in planning, the local, state and central governments have only been partially successful in dealing with these problems. The study on existing policies and programmes also proved that when the government is the sole provider or mediator, GIS can become a tool of coercion rather than participatory decision-making. It has also been observed that local level administrators who have adopted Geospatial technology for local planning continue to base decision-making on existing political processes. In this juncture, geospatial decision support system (GSDSS) can provide a framework for integrating database management systems with analytical models, graphical display, tabular reporting capabilities and the expert knowledge of decision makers. This assists decision-makers to generate and evaluate alternative solutions to spatial problems. During this process, decision-makers undertake a process of decision research - producing a large number of possible decision alternatives and provide opportunities to involve the community in decision making. The objective is to help decision makers and planners to find solutions through a quantitative spatial evaluation and verification process. The study investigates the options for slum development in a formal framework of RAY (Rajiv Awas Yojana), an ambitious program of Indian Government for slum development. The software modules for realizing the GSDSS were developed using the ArcGIS and Community -VIZ software for Gulbarga city.

  15. National Renewable Energy Laboratory information resources catalogue. A collection of energy efficiency and renewable energy information resources

    Energy Technology Data Exchange (ETDEWEB)

    1994-12-31

    NREL`s first annual Information Resources Catalogue is intended to inform anyone interested in energy efficiency and renewable energy technologies of NREL`s outreach activities, including publications and services. For ease of use, all entries are categorized by subject. The catalogue is separated into six main sections. The first section lists and describes services that are available through NREL and how they may be assessed. The second section contains a list of documents that are published by NREL on a regular or periodic basis. The third section highlights NREL`s series publications written for specific audiences and presenting a wide range of subjects. NREL`s General Interest Publications constitute the fourth section of the catalogue and are written for nontechnical audiences. Descriptions are provided for these publications. The fifth section contains Technical Reports that detail research and development projects. The section on Conference Papers/Journal Articles/Book Chapters makes up the sixth and final section of the catalogue.

  16. A web service for service composition to aid geospatial modelers

    Science.gov (United States)

    Bigagli, L.; Santoro, M.; Roncella, R.; Mazzetti, P.

    2012-04-01

    The identification of appropriate mechanisms for process reuse, chaining and composition is considered a key enabler for the effective uptake of a global Earth Observation infrastructure, currently pursued by the international geospatial research community. In the Earth and Space Sciences, such a facility could primarily enable integrated and interoperable modeling, for what several approaches have been proposed and developed, over the last years. In fact, GEOSS is specifically tasked with the development of the so-called "Model Web". At increasing levels of abstraction and generalization, the initial stove-pipe software tools have evolved to community-wide modeling frameworks, to Component-Based Architecture solution, and, more recently, started to embrace Service-Oriented Architectures technologies, such as the OGC WPS specification and the WS-* stack of W3C standards for service composition. However, so far, the level of abstraction seems too low for implementing the Model Web vision, and far too complex technological aspects must still be addressed by both providers and users, resulting in limited usability and, eventually, difficult uptake. As by the recent ICT trend of resource virtualization, it has been suggested that users in need of a particular processing capability, required by a given modeling workflow, may benefit from outsourcing the composition activities into an external first-class service, according to the Composition as a Service (CaaS) approach. A CaaS system provides the necessary interoperability service framework for adaptation, reuse and complementation of existing processing resources (including models and geospatial services in general) in the form of executable workflows. This work introduces the architecture of a CaaS system, as a distributed information system for creating, validating, editing, storing, publishing, and executing geospatial workflows. This way, the users can be freed from the need of a composition infrastructure and

  17. Geospatial Technology In Environmental Impact Assessments – Retrospective.

    Directory of Open Access Journals (Sweden)

    Goparaju Laxmi

    2015-10-01

    Full Text Available Environmental Impact Assessments are studies conducted to give us an insight into the various impacts caused by an upcoming industry or any developmental activity. It should address various social, economic and environmental issues ensuring that negative impacts are mitigated. In this context, geospatial technology has been used widely in recent times.

  18. Integrated web system of geospatial data services for climate research

    Science.gov (United States)

    Okladnikov, Igor; Gordov, Evgeny; Titov, Alexander

    2016-04-01

    Georeferenced datasets are currently actively used for modeling, interpretation and forecasting of climatic and ecosystem changes on different spatial and temporal scales. Due to inherent heterogeneity of environmental datasets as well as their huge size (up to tens terabytes for a single dataset) a special software supporting studies in the climate and environmental change areas is required. An approach for integrated analysis of georefernced climatological data sets based on combination of web and GIS technologies in the framework of spatial data infrastructure paradigm is presented. According to this approach a dedicated data-processing web system for integrated analysis of heterogeneous georeferenced climatological and meteorological data is being developed. It is based on Open Geospatial Consortium (OGC) standards and involves many modern solutions such as object-oriented programming model, modular composition, and JavaScript libraries based on GeoExt library, ExtJS Framework and OpenLayers software. This work is supported by the Ministry of Education and Science of the Russian Federation, Agreement #14.613.21.0037.

  19. Business models for implementing geospatial technologies in transportation decision-making

    Science.gov (United States)

    2007-03-31

    This report describes six State DOTs business models for implementing geospatial technologies. It provides a comparison of the organizational factors influencing how Arizona DOT, Delaware DOT, Georgia DOT, Montana DOT, North Carolina DOT, and Okla...

  20. Urban Image Classification: Per-Pixel Classifiers, Sub-Pixel Analysis, Object-Based Image Analysis, and Geospatial Methods. 10; Chapter

    Science.gov (United States)

    Myint, Soe W.; Mesev, Victor; Quattrochi, Dale; Wentz, Elizabeth A.

    2013-01-01

    Remote sensing methods used to generate base maps to analyze the urban environment rely predominantly on digital sensor data from space-borne platforms. This is due in part from new sources of high spatial resolution data covering the globe, a variety of multispectral and multitemporal sources, sophisticated statistical and geospatial methods, and compatibility with GIS data sources and methods. The goal of this chapter is to review the four groups of classification methods for digital sensor data from space-borne platforms; per-pixel, sub-pixel, object-based (spatial-based), and geospatial methods. Per-pixel methods are widely used methods that classify pixels into distinct categories based solely on the spectral and ancillary information within that pixel. They are used for simple calculations of environmental indices (e.g., NDVI) to sophisticated expert systems to assign urban land covers. Researchers recognize however, that even with the smallest pixel size the spectral information within a pixel is really a combination of multiple urban surfaces. Sub-pixel classification methods therefore aim to statistically quantify the mixture of surfaces to improve overall classification accuracy. While within pixel variations exist, there is also significant evidence that groups of nearby pixels have similar spectral information and therefore belong to the same classification category. Object-oriented methods have emerged that group pixels prior to classification based on spectral similarity and spatial proximity. Classification accuracy using object-based methods show significant success and promise for numerous urban 3 applications. Like the object-oriented methods that recognize the importance of spatial proximity, geospatial methods for urban mapping also utilize neighboring pixels in the classification process. The primary difference though is that geostatistical methods (e.g., spatial autocorrelation methods) are utilized during both the pre- and post

  1. 75 FR 10309 - Announcement of National Geospatial Advisory Committee Meeting

    Science.gov (United States)

    2010-03-05

    ... Geospatial Advisory Committee (NGAC) will meet on March 24-25, 2010 at the One Washington Circle Hotel, 1... implementation of Office of Management and Budget (OMB) Circular A-16. Topics to be addressed at the meeting...

  2. GEO-SPATIAL MODELING OF TRAVEL TIME TO MEDICAL FACILITIES IN MUNA BARAT DISTRICT, SOUTHEAST SULAWESI PROVINCE, INDONESIA

    Directory of Open Access Journals (Sweden)

    Nelson Sula

    2018-03-01

    Full Text Available Background: Health services are strongly influenced by regional topography. Road infrastructure is a key in access to health services. The geographic information system becomes a tool in modeling access to health services. Objective: To analyze geospatial data of the travel time to medical facilities in Muna Barat district, Southeast Sulawesi Province, Indonesia. Methods: This research used geospatial analysis with classification of raster data then overlaid with raster data such as Digital Elevation Modeling (DEM, Road of Vector data, and the point of Public Health Center (Puskesmas. Results: The result of geospatial analysis showed that the travel time to Puskesmas in Napano Kusambi and Kusambi sub districts is between 90-120 minutes, and travel time to the hospital in Kusambi sub district is required more than 2 hours. Conclusion: The output of this geospatial analysis can be an input for local government in planning infrastructure development in Muna Barat District, Indonesia.

  3. IAEA Library Catalogue of Books 1968-1970

    International Nuclear Information System (INIS)

    1971-01-01

    This is the first cumulative volume of the IAEA library new acquisitions. It lists new material received during the period March 1968 - December 1970. The catalogue is divided into four major sections. The first contains the full bibliographic listing for each entry. It is arranged by broad subjects, and within each subject by the Universal Decimal Classification (UDC) number. Each entry was then assigned a consecutive item number. The other three sections contain the personal author, title and corporate entry indexes, respectively

  4. How lives became lists and scientific papers became data: cataloguing authorship during the nineteenth century.

    Science.gov (United States)

    Csiszar, Alex

    2017-03-01

    The Catalogue of Scientific Papers, published by the Royal Society of London beginning in 1867, projected back to the beginning of the nineteenth century a novel vision of the history of science in which knowledge was built up out of discrete papers each connected to an author. Its construction was an act of canon formation that helped naturalize the idea that scientific publishing consisted of special kinds of texts and authors that were set apart from the wider landscape of publishing. By recovering the decisions and struggles through which the Catalogue was assembled, this essay aims to contribute to current efforts to denaturalize the scientific paper as the dominant genre of scientific life. By privileging a specific representation of the course of a scientific life as a list of papers, the Catalogue helped shape underlying assumptions about the most valuable fruits of a scientific career. Its enumerated lists of authors' periodical publications were quickly put to use as a means of measuring scientific productivity and reputation, as well as by writers of biography and history. Although it was first conceived as a search technology, this essay locates the Catalogue's most consequential legacy in its uses as a technology of valuation.

  5. Reviews of Geospatial Information Technology and Collaborative Data Delivery for Disaster Risk Management

    Directory of Open Access Journals (Sweden)

    Hiroyuki Miyazaki

    2015-09-01

    Full Text Available Due to the fact that geospatial information technology is considered necessary for disaster risk management (DRM, the need for more effective collaborations between providers and end users in data delivery is increasing. This paper reviews the following: (i schemes of disaster risk management and collaborative data operation in DRM; (ii geospatial information technology in terms of applications to the schemes reviewed; and (iii ongoing practices of collaborative data delivery with the schemes reviewed. This paper concludes by discussing the future of collaborative data delivery and the progress of the technologies.

  6. Qualitative-Geospatial Methods of Exploring Person-Place Transactions in Aging Adults: A Scoping Review.

    Science.gov (United States)

    Hand, Carri; Huot, Suzanne; Laliberte Rudman, Debbie; Wijekoon, Sachindri

    2017-06-01

    Research exploring how places shape and interact with the lives of aging adults must be grounded in the places where aging adults live and participate. Combined participatory geospatial and qualitative methods have the potential to illuminate the complex processes enacted between person and place to create much-needed knowledge in this area. The purpose of this scoping review was to identify methods that can be used to study person-place relationships among aging adults and their neighborhoods by determining the extent and nature of research with aging adults that combines qualitative methods with participatory geospatial methods. A systematic search of nine databases identified 1,965 articles published from 1995 to late 2015. We extracted data and assessed whether the geospatial and qualitative methods were supported by a specified methodology, the methods of data analysis, and the extent of integration of geospatial and qualitative methods. Fifteen studies were included and used the photovoice method, global positioning system tracking plus interview, or go-along interviews. Most included articles provided sufficient detail about data collection methods, yet limited detail about methodologies supporting the study designs and/or data analysis. Approaches that combine participatory geospatial and qualitative methods are beginning to emerge in the aging literature. By more explicitly grounding studies in a methodology, better integrating different types of data during analysis, and reflecting on methods as they are applied, these methods can be further developed and utilized to provide crucial place-based knowledge that can support aging adults' health, well-being, engagement, and participation. © The Author 2017. Published by Oxford University Press on behalf of The Gerontological Society of America. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  7. Free and Open Source Software for Geospatial in the field of planetary science

    Science.gov (United States)

    Frigeri, A.

    2012-12-01

    Information technology applied to geospatial analyses has spread quickly in the last ten years. The availability of OpenData and data from collaborative mapping projects increased the interest on tools, procedures and methods to handle spatially-related information. Free Open Source Software projects devoted to geospatial data handling are gaining a good success as the use of interoperable formats and protocols allow the user to choose what pipeline of tools and libraries is needed to solve a particular task, adapting the software scene to his specific problem. In particular, the Free Open Source model of development mimics the scientific method very well, and researchers should be naturally encouraged to take part to the development process of these software projects, as this represent a very agile way to interact among several institutions. When it comes to planetary sciences, geospatial Free Open Source Software is gaining a key role in projects that commonly involve different subjects in an international scenario. Very popular software suites for processing scientific mission data (for example, ISIS) and for navigation/planning (SPICE) are being distributed along with the source code and the interaction between user and developer is often very strict, creating a continuum between these two figures. A very widely spread library for handling geospatial data (GDAL) has started to support planetary data from the Planetary Data System, and recent contributions enabled the support to other popular data formats used in planetary science, as the Vicar one. The use of Geographic Information System in planetary science is now diffused, and Free Open Source GIS, open GIS formats and network protocols allow to extend existing tools and methods developed to solve Earth based problems, also to the case of the study of solar system bodies. A day in the working life of a researcher using Free Open Source Software for geospatial will be presented, as well as benefits and

  8. New displaying models of bibliographic data and resources: cataloguing/resource description and search results

    Directory of Open Access Journals (Sweden)

    Antonella Trombone

    2014-05-01

    The paper here proposed analyzes different patterns of bibliographic data visualization that libraries’ OPACs and library service platforms begin to offer, focusing on three categories of innovations in displaying of cataloguing data: the data proposed by discovery tools or library service platforms, that transform and integrate information taken from the Integrated Library Systems with other collections accessible through libraries; the data derived from the cataloguing based on RDA started in some libraries; the displays of bibliographic data emulating the hierarchical organization among entities foreseen by FRBR.

  9. Investigating Climate Change Issues With Web-Based Geospatial Inquiry Activities

    Science.gov (United States)

    Dempsey, C.; Bodzin, A. M.; Sahagian, D. L.; Anastasio, D. J.; Peffer, T.; Cirucci, L.

    2011-12-01

    In the Environmental Literacy and Inquiry middle school Climate Change curriculum we focus on essential climate literacy principles with an emphasis on weather and climate, Earth system energy balance, greenhouse gases, paleoclimatology, and how human activities influence climate change (http://www.ei.lehigh.edu/eli/cc/). It incorporates a related set of a framework and design principles to provide guidance for the development of the geospatial technology-integrated Earth and environmental science curriculum materials. Students use virtual globes, Web-based tools including an interactive carbon calculator and geologic timeline, and inquiry-based lab activities to investigate climate change topics. The curriculum includes educative curriculum materials that are designed to promote and support teachers' learning of important climate change content and issues, geospatial pedagogical content knowledge, and geographic spatial thinking. The curriculum includes baseline instructional guidance for teachers and provides implementation and adaptation guidance for teaching with diverse learners including low-level readers, English language learners and students with disabilities. In the curriculum, students use geospatial technology tools including Google Earth with embedded spatial data to investigate global temperature changes, areas affected by climate change, evidence of climate change, and the effects of sea level rise on the existing landscape. We conducted a designed-based research implementation study with urban middle school students. Findings showed that the use of the Climate Change curriculum showed significant improvement in urban middle school students' understanding of climate change concepts.

  10. International Atomic Energy Agency publications. Publications catalogue 2004

    International Nuclear Information System (INIS)

    2004-03-01

    This Publications Catalogue lists all sales publications of the IAEA published in 2002, 2003 and forthcoming in early 2004. Most IAEA publications are issued in English, though some are also available in Arabic, Chinese, French, Russian or Spanish. This is indicated at the bottom of the book entry. A complete listing of all IAEA priced publications is available on the IAEA's web site: http://www.iaea.org/books

  11. Organizational needs for managing and preserving geospatial data and related electronic records

    Directory of Open Access Journals (Sweden)

    R R Downs

    2006-01-01

    Full Text Available Government agencies and other organizations are required to manage and preserve records that they create and use to facilitate future access and reuse. The increasing use of geospatial data and related electronic records presents new challenges for these organizations, which have relied on traditional practices for managing and preserving records in printed form. This article reports on an investigation of current and future needs for managing and preserving geospatial electronic records on the part of localand state-level organizations in the New York City metropolitan region. It introduces the study and describes organizational needs observed, including needs for organizational coordination and interorganizational cooperation throughout the entire data lifecycle.

  12. Geospatial Health: the first five years

    Directory of Open Access Journals (Sweden)

    Jürg Utzinger

    2011-11-01

    Full Text Available Geospatial Health is an international, peer-reviewed scientific journal produced by the Global Network for Geospatial Health (GnosisGIS. This network was founded in 2000 and the inaugural issue of its official journal was published in November 2006 with the aim to cover all aspects of geographical information system (GIS applications, remote sensing and other spatial analytic tools focusing on human and veterinary health. The University of Naples Federico II is the publisher, producing two issues per year, both as hard copy and an open-access online version. The journal is referenced in major databases, including CABI, ISI Web of Knowledge and PubMed. In 2008, it was assigned its first impact factor (1.47, which has now reached 1.71. Geospatial Health is managed by an editor-in-chief and two associate editors, supported by five regional editors and a 23-member strong editorial board. This overview takes stock of the first five years of publishing: 133 contributions have been published so far, primarily original research (79.7%, followed by reviews (7.5%, announcements (6.0%, editorials and meeting reports (3.0% each and a preface in the first issue. A content analysis of all the original research articles and reviews reveals that three quarters of the publications focus on human health with the remainder dealing with veterinary health. Two thirds of the papers come from Africa, Asia and Europe with similar numbers of contributions from each continent. Studies of more than 35 different diseases, injuries and risk factors have been presented. Malaria and schistosomiasis were identified as the two most important diseases (11.2% each. Almost half the contributions were based on GIS, one third on spatial analysis, often using advanced Bayesian geostatistics (13.8%, and one quarter on remote sensing. The 120 original research articles, reviews and editorials were produced by 505 authors based at institutions and universities in 52 countries

  13. FOSS geospatial libraries in scientific workflow environments: experiences and directions

    CSIR Research Space (South Africa)

    McFerren, G

    2011-07-01

    Full Text Available of experiments. In context of three sets of research (wildfire research, flood modelling and the linking of disease outbreaks to multi-scale environmental conditions), we describe our efforts to provide geospatial capability for scientific workflow software...

  14. A new catalogue of earthquakes in the historical Armenian area from antiquity to the 12th century

    Directory of Open Access Journals (Sweden)

    G. Traina

    1995-06-01

    Full Text Available The present contribution describes the method of work, the types of source materia] used, and the historio- graphical and historico-eismic tradition of Armenia. The catalogue' s territorial frame of reference is that of socalled historical Armenia (which included part of present Eastern Turkey, and part of present Azerbaijan. The sources belong to different languages and cultures: Armenian, Syriac, Greek, Arab, Persian and Georgian. A comparison of the local sources with those belonging to other cultures enab]es the historical and seismological I"adition of the Mediterl'anean to be "linked" with that of the Iranian p]ateau, traditionally considered as two separate areas. We analyzed historical events listed in the most recent catalogues of earthquakes in the Armenian area compiled by Kondorskaya and Shebalin (1982 and Karapetian (1991. Important and valuable though these catalogues are, they are in need of revision. We found evidence for six hitherto unrecorded seismic events. Numerous errors of dating and location have been corrected, and several new localities and seismic effects have been evidenced. Each modification of the previous catalogues has been documented on the hasis of the historiographical and literary sources and the data from the written sources have been linked with those concerning the history of Armenian cities and architecture (monasteries, churches, episcopal complexes. On the whole. the revised earthquakes seem underestimated in the previous catalogues. The aim of this catalogue is to make a contribution to the knowledge of historical seismicity in Armenia, and at the same time to underline the specific nature of the Armenian case, thus avoiding a procedure which has generally tended to place this area in a marginal position, within the wider field of other research on historical earthquakes.

  15. Learning transfer of geospatial technologies in secondary science and mathematics core areas

    Science.gov (United States)

    Nielsen, Curtis P.

    The purpose of this study was to investigate the transfer of geospatial technology knowledge and skill presented in a social sciences course context to other core areas of the curriculum. Specifically, this study explored the transfer of geospatial technology knowledge and skill to the STEM-related core areas of science and mathematics among ninth-grade students. Haskell's (2001) research on "levels of transfer" provided the theoretical framework for this study, which sought to demonstrate the experimental group's higher ability to transfer geospatial skills, higher mean assignment scores, higher post-test scores, higher geospatial skill application and deeper levels of transfer application than the control group. The participants of the study consisted of thirty ninth-graders enrolled in U.S. History, Earth Science and Integrated Mathematics 1 courses. The primary investigator of this study had no previous classroom experiences with this group of students. The participants who were enrolled in the school's existing two-section class configuration were assigned to experimental and control groups. The experimental group had ready access to Macintosh MacBook laptop computers, and the control group had ready access to Macintosh iPads. All participants in U.S. History received instruction with and were required to use ArcGIS Explorer Online during a Westward Expansion project. All participants were given the ArcGIS Explorer Online content assessment following the completion of the U.S. History project. Once the project in U.S. History was completed, Earth Science and Integrated Mathematics 1 began units of instruction beginning with a multiple-choice content pre-test created by the classroom teachers. Experimental participants received the same unit of instruction without the use or influence of ArcGIS Explorer Online. At the end of the Earth Science and Integrated Math 1 units, the same multiple-choice test was administered as the content post-test. Following the

  16. The Impact of a Geospatial Technology-Supported Energy Curriculum on Middle School Students' Science Achievement

    Science.gov (United States)

    Kulo, Violet; Bodzin, Alec

    2013-02-01

    Geospatial technologies are increasingly being integrated in science classrooms to foster learning. This study examined whether a Web-enhanced science inquiry curriculum supported by geospatial technologies promoted urban middle school students' understanding of energy concepts. The participants included one science teacher and 108 eighth-grade students classified in three ability level tracks. Data were gathered through pre/posttest content knowledge assessments, daily classroom observations, and daily reflective meetings with the teacher. Findings indicated a significant increase in the energy content knowledge for all the students. Effect sizes were large for all three ability level tracks, with the middle and low track classes having larger effect sizes than the upper track class. Learners in all three tracks were highly engaged with the curriculum. Curriculum effectiveness and practical issues involved with using geospatial technologies to support science learning are discussed.

  17. Geospatial Analysis of Low-frequency Radio Signals Collected During the 2017 Solar Eclipse

    Science.gov (United States)

    Liles, W. C.; Nelson, J.; Kerby, K. C.; Lukes, L.; Henry, J.; Oputa, J.; Lemaster, G.

    2017-12-01

    The total solar eclipse of 2017, with a path that crosses the continental United States, offers a unique opportunity to gather geospatially diverse data. The EclipseMob project has been designed to crowdsource this data by building a network of citizen scientists across the country. The project focuses on gathering low-frequency radio wave data before, during, and after the eclipse. WWVB, a 60 KHz transmitter in Ft. Collins, CO operated by the National Institutes of Standard and Technology, will provide the transmit signal that will be observed by project participants. Participating citizen scientists are building simple antennas and receivers designed by the EclipseMob team and provided to participants in the form of "receiver kits." The EclipseMob receiver downsamples the 60 KHz signal to 18 KHz and supplies the downsampled signal to the audio jack of a smartphone. A dedicated app is used to collect data and upload it to the EclipseMob server. By studying the variations in WWVB amplitude observed during the eclipse at over 150 locations across the country, we aim to understand how the ionization of the D layer of the ionosphere is impacted by the eclipse as a function of both time and space (location). The diverse locations of the EclipseMob participants will provide data from a wide variety of propagation paths - some crossing the path of the total eclipse, and some remaining on the same side of the eclipse path as the transmitter. Our initial data analysis will involve identifying characteristics that define geospatial relationships in the behavior of observed WWVB signal amplitudes.

  18. Design and development of correlation techniques to maintain a space surveillance system catalogue

    Science.gov (United States)

    Olmedo, E.; Sánchez Ortiz, Noelia; Lerate, Mercedes; Belló-Mora, Miguel; Klinkrad, H.

    2009-10-01

    A growing interest exists in a future, autonomous European Space Surveillance System (ESSS). Currently, most of the knowledge about Earth-orbiting space objects is based on information provided by the USASPACECOM. This paper presents the required initial orbit determination (IOD) and correlation techniques to process optical measurements. Former studies were focused on the handling of radar measurements, which are summarised with the aim of describing a global procedure for processing hybrid measurement types (combination of radar and optic data for catalogue maintenance). The introduction of manoeuvres are presented due to their importance in the space object catalogue maintenance. The detection of uncatalogued objects and the successful correlation of already catalogued objects involve two different tasks for telescopes: survey and tasking. Assumptions for both strategies are developed on the basis of the previous work developed at the University of Berne (see [T. Flohrer, T. Schildknecht, R. Musci, E. Stöveken, Performance estimation for GEO space surveillance, Advances in Space Research 35 (2005). [1]; T. Flohrer, T. Schildknecht, R. Musci, Proposed strategies for optical observations in a future European Space Surveillance Network, presented in the 36th COSPAR Scientific Assembly (2006). [2]; R. Musci, T. Schildknecht, M. Ploner, Orbit improvement for GEO objects using follow-up observations, Advances in Space Research 34 (2004). [3]; R. Musci, T. Schildknecht, M. Ploner, G. Beutler, Orbit improvement for GTO objects using follow-up observations, Advances in Space Research 35 (2005). [4]; R. Musci, T. Schildknecht, T. Flohrer, G. Beutler, Concept for a catalogue of space debris in GEO, Proceedings of the Fourth European Conference on Space Debris, (ESA SP-587, 2005). [5

  19. A multimembership catalogue for 1876 open clusters using UCAC4 data

    Science.gov (United States)

    Sampedro, L.; Dias, W. S.; Alfaro, E. J.; Monteiro, H.; Molino, A.

    2017-10-01

    The main objective of this work is to determine the cluster members of 1876 open clusters, using positions and proper motions of the astrometric fourth United States Naval Observatory (USNO) CCD Astrograph Catalog (UCAC4). For this purpose, we apply three different methods, all based on a Bayesian approach, but with different formulations: a purely parametric method, another completely non-parametric algorithm and a third, recently developed by Sampedro & Alfaro, using both formulations at different steps of the whole process. The first and second statistical moments of the members' phase-space subspace, obtained after applying the three methods, are compared for every cluster. Although, on average, the three methods yield similar results, there are also specific differences between them, as well as for some particular clusters. The comparison with other published catalogues shows good agreement. We have also estimated, for the first time, the mean proper motion for a sample of 18 clusters. The results are organized in a single catalogue formed by two main files, one with the most relevant information for each cluster, partially including that in UCAC4, and the other showing the individual membership probabilities for each star in the cluster area. The final catalogue, with an interface design that enables an easy interaction with the user, is available in electronic format at the Stellar Systems Group (SSG-IAA) web site (http://ssg.iaa.es/en/content/sampedro-cluster-catalog).

  20. Acquisition and Cataloguing Processes: Changes as a Result of Customer Value Discovery Research

    Directory of Open Access Journals (Sweden)

    Sue McKnight

    2007-12-01

    Full Text Available Objective ‐ This study seeks to highlight the profound effect of Customer Value Discovery research on the internal business processes of two university libraries in the areas of cataloguing and acquisitions.Methods ‐ In this project, “Customer Discovery Workshops” with academic staff, students, and university stakeholders provided library managers and staff with information on what services and resources were of value to customers. The workshops also aimed to discover what features of existing library services and resources irritated the students, staff, and faculty. A student satisfaction survey assessed longer‐term impact of library changes to students in one university.Results ‐ The findings resulted in significant changes to collection development, acquisitions, and cataloguing processes. A number of value added services were introduced for the customer. The project also resulted in greater speed and efficiency in dealing with collection development, acquisitions, and cataloguing by the introduction of more technology‐enhanced services. Overall customer satisfaction was improved during the project period.Conclusion ‐ The changes to services introduced as a result of customer feedback also improved relationships between librarians and their university community, through the introduction of a more proactive and supportive service.

  1. A JPEG backward-compatible HDR image compression

    Science.gov (United States)

    Korshunov, Pavel; Ebrahimi, Touradj

    2012-10-01

    High Dynamic Range (HDR) imaging is expected to become one of the technologies that could shape next generation of consumer digital photography. Manufacturers are rolling out cameras and displays capable of capturing and rendering HDR images. The popularity and full public adoption of HDR content is however hindered by the lack of standards in evaluation of quality, file formats, and compression, as well as large legacy base of Low Dynamic Range (LDR) displays that are unable to render HDR. To facilitate wide spread of HDR usage, the backward compatibility of HDR technology with commonly used legacy image storage, rendering, and compression is necessary. Although many tone-mapping algorithms were developed for generating viewable LDR images from HDR content, there is no consensus on which algorithm to use and under which conditions. This paper, via a series of subjective evaluations, demonstrates the dependency of perceived quality of the tone-mapped LDR images on environmental parameters and image content. Based on the results of subjective tests, it proposes to extend JPEG file format, as the most popular image format, in a backward compatible manner to also deal with HDR pictures. To this end, the paper provides an architecture to achieve such backward compatibility with JPEG and demonstrates efficiency of a simple implementation of this framework when compared to the state of the art HDR image compression.

  2. THROES: a caTalogue of HeRschel Observations of Evolved Stars. I. PACS range spectroscopy

    Science.gov (United States)

    Ramos-Medina, J.; Sánchez Contreras, C.; García-Lario, P.; Rodrigo, C.; da Silva Santos, J.; Solano, E.

    2018-03-01

    This is the first of a series of papers presenting the THROES (A caTalogue of HeRschel Observations of Evolved Stars) project, intended to provide a comprehensive overview of the spectroscopic results obtained in the far-infrared (55-670 μm) with the Herschel space observatory on low-to-intermediate mass evolved stars in our Galaxy. Here we introduce the catalogue of interactively reprocessed Photoconductor Array Camera and Spectrometer (PACS) spectra covering the 55-200 μm range for 114 stars in this category for which PACS range spectroscopic data is available in the Herschel Science Archive (HSA). Our sample includes objects spanning a range of evolutionary stages, from the asymptotic giant branch to the planetary nebula phase, displaying a wide variety of chemical and physical properties. The THROES/PACS catalogue is accessible via a dedicated web-based interface and includes not only the science-ready Herschel spectroscopic data for each source, but also complementary photometric and spectroscopic data from other infrared observatories, namely IRAS, ISO, or AKARI, at overlapping wavelengths. Our goal is to create a legacy-value Herschel dataset that can be used by the scientific community in the future to deepen our knowledge and understanding of these latest stages of the evolution of low-to-intermediate mass stars. The THROES/PACS catalogue is accessible at http://https://throes.cab.inta-csic.es/

  3. High performance geospatial and climate data visualization using GeoJS

    Science.gov (United States)

    Chaudhary, A.; Beezley, J. D.

    2015-12-01

    GeoJS (https://github.com/OpenGeoscience/geojs) is an open-source library developed to support interactive scientific and geospatial visualization of climate and earth science datasets in a web environment. GeoJS has a convenient application programming interface (API) that enables users to harness the fast performance of WebGL and Canvas 2D APIs with sophisticated Scalable Vector Graphics (SVG) features in a consistent and convenient manner. We started the project in response to the need for an open-source JavaScript library that can combine traditional geographic information systems (GIS) and scientific visualization on the web. Many libraries, some of which are open source, support mapping or other GIS capabilities, but lack the features required to visualize scientific and other geospatial datasets. For instance, such libraries are not be capable of rendering climate plots from NetCDF files, and some libraries are limited in regards to geoinformatics (infovis in a geospatial environment). While libraries such as d3.js are extremely powerful for these kinds of plots, in order to integrate them into other GIS libraries, the construction of geoinformatics visualizations must be completed manually and separately, or the code must somehow be mixed in an unintuitive way.We developed GeoJS with the following motivations:• To create an open-source geovisualization and GIS library that combines scientific visualization with GIS and informatics• To develop an extensible library that can combine data from multiple sources and render them using multiple backends• To build a library that works well with existing scientific visualizations tools such as VTKWe have successfully deployed GeoJS-based applications for multiple domains across various projects. The ClimatePipes project funded by the Department of Energy, for example, used GeoJS to visualize NetCDF datasets from climate data archives. Other projects built visualizations using GeoJS for interactively exploring

  4. Dynamical problems in the preparation of the catalogue of one-apparition comets

    International Nuclear Information System (INIS)

    Sitarski, G.

    1979-01-01

    The idea of recomputing all the orbits of long-period comets arose about ten years ago. The work was undertaken at the Warsaw Astronomical Observatory with the cooperation of Slovakian astronomers at the Astronomical Institute in Bratislava and Tatranska Lomnica. To make reasonable the idea of the new catalogue, the following problems have been taken into consideration when recomputing the cometary orbits: 1. to collect all the observations of one-apparition comets and to reduce them to the one system of star catalogues; 2. to define precisely various types of observations and to include in them the corrections of precession, aberration, etc.; 3. to determine the mathematical criteria for eliminating and weighting the observations; and 4. to take into account nongravitational effects in the comet's motion. (Auth.)

  5. Binary catalogue of exoplanets

    Science.gov (United States)

    Schwarz, Richard; Bazso, Akos; Zechner, Renate; Funk, Barbara

    2016-02-01

    Since 1995 there is a database which list most of the known exoplanets (The Extrasolar Planets Encyclopaedia at http://exoplanet.eu/). With the growing number of detected exoplanets in binary and multiple star systems it became more important to mark and to separate them into a new database, which is not available in the Extrasolar Planets Encyclopaedia. Therefore we established an online database (which can be found at: http://www.univie.ac.at/adg/schwarz/multiple.html) for all known exoplanets in binary star systems and in addition for multiple star systems, which will be updated regularly and linked to the Extrasolar Planets Encyclopaedia. The binary catalogue of exoplanets is available online as data file and can be used for statistical purposes. Our database is divided into two parts: the data of the stars and the planets, given in a separate list. We describe also the different parameters of the exoplanetary systems and present some applications.

  6. Geospatial environmental data modelling applications using remote sensing, GIS and spatial statistics

    Energy Technology Data Exchange (ETDEWEB)

    Siljander, M.

    2010-07-01

    This thesis presents novel modelling applications for environmental geospatial data using remote sensing, GIS and statistical modelling techniques. The studied themes can be classified into four main themes: (i) to develop advanced geospatial databases. Paper (I) demonstrates the creation of a geospatial database for the Glanville fritillary butterfly (Melitaea cinxia) in the Aaland Islands, south-western Finland; (ii) to analyse species diversity and distribution using GIS techniques. Paper (II) presents a diversity and geographical distribution analysis for Scopulini moths at a world-wide scale; (iii) to study spatiotemporal forest cover change. Paper (III) presents a study of exotic and indigenous tree cover change detection in Taita Hills Kenya using airborne imagery and GIS analysis techniques; (iv) to explore predictive modelling techniques using geospatial data. In Paper (IV) human population occurrence and abundance in the Taita Hills highlands was predicted using the generalized additive modelling (GAM) technique. Paper (V) presents techniques to enhance fire prediction and burned area estimation at a regional scale in East Caprivi Namibia. Paper (VI) compares eight state-of-the-art predictive modelling methods to improve fire prediction, burned area estimation and fire risk mapping in East Caprivi Namibia. The results in Paper (I) showed that geospatial data can be managed effectively using advanced relational database management systems. Metapopulation data for Melitaea cinxia butterfly was successfully combined with GPS-delimited habitat patch information and climatic data. Using the geospatial database, spatial analyses were successfully conducted at habitat patch level or at more coarse analysis scales. Moreover, this study showed it appears evident that at a large-scale spatially correlated weather conditions are one of the primary causes of spatially correlated changes in Melitaea cinxia population sizes. In Paper (II) spatiotemporal characteristics

  7. The Road to Responsive: University of Toronto Libraries’ Journey to a New Library Catalogue Interface

    Directory of Open Access Journals (Sweden)

    Lisa Gayhart

    2014-01-01

    Full Text Available With the recent surge in the mobile device market and an ever expanding patron base with increasingly divergent levels of technical ability, the University of Toronto Libraries embarked on the development of a new catalogue discovery layer to fit the needs of its diverse users. The result: a mobile-friendly, flexible and intuitive web application that brings the full power of a faceted library catalogue to users without compromising quality or performance, employing Responsive Web Design principles.

  8. Parallel Agent-as-a-Service (P-AaaS Based Geospatial Service in the Cloud

    Directory of Open Access Journals (Sweden)

    Xicheng Tan

    2017-04-01

    Full Text Available To optimize the efficiency of the geospatial service in the flood response decision making system, a Parallel Agent-as-a-Service (P-AaaS method is proposed and implemented in the cloud. The prototype system and comparisons demonstrate the advantages of our approach over existing methods. The P-AaaS method includes both parallel architecture and a mechanism for adjusting the computational resources—the parallel geocomputing mechanism of the P-AaaS method used to execute a geospatial service and the execution algorithm of the P-AaaS based geospatial service chain, respectively. The P-AaaS based method has the following merits: (1 it inherits the advantages of the AaaS-based method (i.e., avoiding transfer of large volumes of remote sensing data or raster terrain data, agent migration, and intelligent conversion into services to improve domain expert collaboration; (2 it optimizes the low performance and the concurrent geoprocessing capability of the AaaS-based method, which is critical for special applications (e.g., highly concurrent applications and emergency response applications; and (3 it adjusts the computing resources dynamically according to the number and the performance requirements of concurrent requests, which allows the geospatial service chain to support a large number of concurrent requests by scaling up the cloud-based clusters in use and optimizes computing resources and costs by reducing the number of virtual machines (VMs when the number of requests decreases.

  9. Geospatial Image Mining For Nuclear Proliferation Detection: Challenges and New Opportunities

    Energy Technology Data Exchange (ETDEWEB)

    Vatsavai, Raju [ORNL; Bhaduri, Budhendra L [ORNL; Cheriyadat, Anil M [ORNL; Arrowood, Lloyd [Y-12 National Security Complex; Bright, Eddie A [ORNL; Gleason, Shaun Scott [ORNL; Diegert, Carl [Sandia National Laboratories (SNL); Katsaggelos, Aggelos K [ORNL; Pappas, Thrasos N [ORNL; Porter, Reid [Los Alamos National Laboratory (LANL); Bollinger, Jim [Savannah River National Laboratory (SRNL); Chen, Barry [Lawrence Livermore National Laboratory (LLNL); Hohimer, Ryan [Pacific Northwest National Laboratory (PNNL)

    2010-01-01

    With increasing understanding and availability of nuclear technologies, and increasing persuasion of nuclear technologies by several new countries, it is increasingly becoming important to monitor the nuclear proliferation activities. There is a great need for developing technologies to automatically or semi-automatically detect nuclear proliferation activities using remote sensing. Images acquired from earth observation satellites is an important source of information in detecting proliferation activities. High-resolution remote sensing images are highly useful in verifying the correctness, as well as completeness of any nuclear program. DOE national laboratories are interested in detecting nuclear proliferation by developing advanced geospatial image mining algorithms. In this paper we describe the current understanding of geospatial image mining techniques and enumerate key gaps and identify future research needs in the context of nuclear proliferation.

  10. Contextualizing Cave Maps as Geospatial Information: Case Study of Indonesia

    Science.gov (United States)

    Reinhart, H.

    2017-12-01

    Caves are the result of solution processes. Because they are happened from geochemical and tectonic activity, they can be considered as geosphere phenomena. As one of the geosphere phenomena, especially at karst landform, caves have spatial dimensions and aspects. Cave’s utilizations and developments are increasing in many sectors such as hydrology, earth science, and tourism industry. However, spatial aspects of caves are poorly concerned dues to the lack of recognition toward cave maps. Many stakeholders have not known significances and importance of cave maps in determining development of a cave. Less information can be considered as the cause. Therefore, it is strongly necessary to put cave maps into the right context in order to make stakeholders realize the significance of it. Also, cave maps will be officially regarded as tools related to policy, development, and conservation act of caves hence they will have regulation in the usages and applications. This paper aims to make the contextualization of cave maps toward legal act. The act which is used is Act Number 4 Year 2011 About Geospatial Information. The contextualization is done by scrutinizing every articles and clauses related to cave maps and seek the contextual elements from both of them. The results are that cave maps can be regarded as geospatial information and classified as thematic geospatial information. The usages of them can be regulated through the Act Number 4 Year 2011. The regulations comprised by data acquisition, database, authorities, surveyor, and the obligation of providing cave maps in planning cave’s development and the environment surrounding.

  11. Geo-Spatial Tactical Decision Aid Systems: Fuzzy Logic for Supporting Decision Making

    National Research Council Canada - National Science Library

    Grasso, Raffaele; Giannecchini, Simone

    2006-01-01

    .... This paper describes a tactical decision aid system based on fuzzy logic reasoning for data fusion and on current Open Geospatial Consortium specifications for interoperability, data dissemination...

  12. Marine chronometers at Greenwich a catalogue of marine chronometers at the National Maritime Museum, Greenwich

    CERN Document Server

    Betts, Jonathan

    2018-01-01

    This is a comprehensive, illustrated catalogue of the 200+ marine chronometers in the collections of Royal Museums Greenwich. Every chronometer has been completely dismantled, studied and recorded, and illustrations include especially commissioned line drawings as well as photographs. The collection is also used to illustrate a newly researched and up-to-date chapter describing the history of the marine chronometer, so the book is much more than simply a catalogue. The history chapter naturally includes the story of John Harrison’s pioneering work in creating the first practical marine timekeepers, all four of which are included in the catalogue, newly photographed and described in minute detail for the first time. In fact full technical and historical data are provided for all of the marine chronometers in the collection, to an extent never before attempted, including biographical details of every maker represented. A chapter describes how the 19th century English chronometer was manufactured, and another ...

  13. Information gathering, management and transferring for geospatial intelligence - A conceptual approach to create a spatial data infrastructure

    Science.gov (United States)

    Nunes, Paulo; Correia, Anacleto; Teodoro, M. Filomena

    2017-06-01

    Since long ago, information is a key factor for military organizations. In military context the success of joint and combined operations depends on the accurate information and knowledge flow concerning the operational theatre: provision of resources, environment evolution, targets' location, where and when an event will occur. Modern military operations cannot be conceive without maps and geospatial information. Staffs and forces on the field request large volume of information during the planning and execution process, horizontal and vertical geospatial information integration is critical for decision cycle. Information and knowledge management are fundamental to clarify an environment full of uncertainty. Geospatial information (GI) management rises as a branch of information and knowledge management, responsible for the conversion process from raw data collect by human or electronic sensors to knowledge. Geospatial information and intelligence systems allow us to integrate all other forms of intelligence and act as a main platform to process and display geospatial-time referenced events. Combining explicit knowledge with person know-how to generate a continuous learning cycle that supports real time decisions, mitigates the influences of fog of war and provides the knowledge supremacy. This paper presents the analysis done after applying a questionnaire and interviews about the GI and intelligence management in a military organization. The study intended to identify the stakeholder's requirements for a military spatial data infrastructure as well as the requirements for a future software system development.

  14. Geospatial Visualization of Scientific Data Through Keyhole Markup Language

    Science.gov (United States)

    Wernecke, J.; Bailey, J. E.

    2008-12-01

    The development of virtual globes has provided a fun and innovative tool for exploring the surface of the Earth. However, it has been the paralleling maturation of Keyhole Markup Language (KML) that has created a new medium and perspective through which to visualize scientific datasets. Originally created by Keyhole Inc., and then acquired by Google in 2004, in 2007 KML was given over to the Open Geospatial Consortium (OGC). It became an OGC international standard on 14 April 2008, and has subsequently been adopted by all major geobrowser developers (e.g., Google, Microsoft, ESRI, NASA) and many smaller ones (e.g., Earthbrowser). By making KML a standard at a relatively young stage in its evolution, developers of the language are seeking to avoid the issues that plagued the early World Wide Web and development of Hypertext Markup Language (HTML). The popularity and utility of Google Earth, in particular, has been enhanced by KML features such as the Smithsonian volcano layer and the dynamic weather layers. Through KML, users can view real-time earthquake locations (USGS), view animations of polar sea-ice coverage (NSIDC), or read about the daily activities of chimpanzees (Jane Goodall Institute). Perhaps even more powerful is the fact that any users can create, edit, and share their own KML, with no or relatively little knowledge of manipulating computer code. We present an overview of the best current scientific uses of KML and a guide to how scientists can learn to use KML themselves.

  15. Planck early results. XIII. Statistical properties of extragalactic radio sources in the Planck Early Release Compact Source Catalogue

    DEFF Research Database (Denmark)

    Lähteenmäki, A.; Poutanen, T.; Natoli, P.

    2011-01-01

    The data reported in Planck's Early Release Compact Source Catalogue (ERCSC) are exploited to measure the number counts (dN/dS) of extragalactic radio sources at 30, 44, 70, 100, 143 and 217 GHz. Due to the full-sky nature of the catalogue, this measurement extends to the rarest and brightest sou...

  16. VizieR Online Data Catalog: Catalogue of Radio Stars (Wendker, 2001)

    Science.gov (United States)

    Wendker, H. J.

    2015-06-01

    The first version of this catalogue was published in Abh.Hamburger Sternw. 1978, Vol.10, p 1ff. (CDS Catalogue II/129). A second version was published in 1987 (1987A&AS...69...87W) and microfiches (CDS Catalogue II/147). A third version was published 1995A&AS..109..177W (CDS Catalogue II/199). The basic concept of the earlier versions is preserved (in file "catalog.txt"), namely one entry per star per frequency per paper. Space is now provided, however, to add more informations. These may be of technical or astronomical nature. Usually month and year of observation and the number of independent data points or length of monitoring session are given. In the file "catalog.txt", all radio data are preceded by a header which contains information on the star or stellar system. (Note, that a physical stellar system is regarded as one single entry and that comments pertaining to individual components are found directly behind the observational data). Stellar data like names, position, proper motion, magnitudes and spectroscopic types are given in fixed format in a self-explanatory fashion. It is tried to have typical values from commonly available references. It is not intended to compete here with other compilations. These header informations are collected when the star is entered for the first time. They are only changed when new values are available while additional radio references are added. An arbitrarily expandable section for unformatted text finishes the header. Finally, the units of the radio data remain in MHz (column#1) and mJy (columns #2. and #3). All coordinates refer to epoch and equinox 1950.0 (e.g. B1950). This is a so-called merged version e.g. all stars, those detected at least once and those with upper limits only, are listed in order of ascending right ascension. The detected stars are marked with a "D" in the outermost right hand column in lines 1 to 5 ('D' in column "Det" of the file "stars.dat"). The last updating occurred on 2001-Mar-06. In this

  17. Nuclear standards

    International Nuclear Information System (INIS)

    Fichtner, N.; Becker, K.; Bashir, M.

    1981-01-01

    This compilation of all nuclear standards available to the authors by mid 1980 represents the third, carefully revised edition of a catalogue which was first published in 1975 as EUR 5362. In this third edition several changes have been made. The title has been condensed. The information has again been carefully up-dated, covering all changes regarding status, withdrawal of old standards, new projects, amendments, revisions, splitting of standards into several parts, combination of several standards into one, etc., as available to the authors by mid 1980. The speed with which information travels varies and requires in many cases rather tedious and cumbersome inquiries. Also, the classification scheme has been revised with the goal of better adjustment to changing situations and priorities. Whenever it turned out to be difficult to attribute a standard to a single subject category, multiple listings in all relevant categories have been made. As in previous editions, within the subcategories the standards are arranged by organization (in Categorie 2.1 by country) alphabetically and in ascending numerical order. It covers all relevant areas of power reactors, the fuel cycle, radiation protection, etc., from the basic laws and governmental regulations, regulatory guides, etc., all the way to voluntary industrial standards and codes of pratice. (orig./HP)

  18. False become true: Christie's catalogue of Zhang Hongtu

    Directory of Open Access Journals (Sweden)

    Sandra Valenzuela Arellano

    2013-06-01

    Full Text Available By analyzing the series Christie’s Catalogue Project by the artist Zhang Hongtu, this essay traces relationships between: Cultural heritance, national identity, fakeness, popular culture and legitimacy of power structures, within contemporary China. Contemporary art creates assemblies of meanings related to its time, sometimes relating the present with history, high with vernacular culture, soft power with the construction of collective memory. This essay aims to reconstruct with words, through textual and visual analysis, those visual assemblies.

  19. Analysis of the presence of pseudoscience in the catalogues of public libraries in Spain

    Directory of Open Access Journals (Sweden)

    Sergi Cortiñas-Rovira

    2018-02-01

    Full Text Available This paper examines the presence of books regarding instances of pseudoscience as their main topic in catalogues of public libraries in Spain, as well as the opinion of such works about pseudoscience. A database is created to analyse both absolute and relative amounts of those works in all Spanish catalogues. The numbers show a small, yet worrying percentage of 0.15% titles concerning several pseudosciences and prove astrology as the most abundant instance in libraries. The study also explains the menace of homeopathy, as the works about this pseudotherapy outnumber those regarding chemotherapy in a 1/20 ratio.

  20. Catalogue of standards at the Safety Analysis department library, the first of March 1987

    International Nuclear Information System (INIS)

    1987-01-01

    This report is a compilation of ANS and ANSI/ANS, ANSI/ASME american standards, Regulatory guides (Power Reactor Division), IEEE standards, to which are added ANSI american standards, AFNOR french standards, all of them being updated until to day. This computed compilation is divided into three parts: 1) main descriptor categories; 2) list of american standards in chronological order, list of AFNOR french standards, list of Regulatory guides in numerical order. 3) list of descriptors in alphabetic order [fr

  1. Geospatial Associations Between Tobacco Retail Outlets and Current Use of Cigarettes and e-Cigarettes among Youths in Texas.

    Science.gov (United States)

    Pérez, Adriana; Chien, Lung-Chang; Harrell, Melissa B; Pasch, Keryn E; Obinwa, Udoka C; Perry, Cheryl L

    2017-10-01

    To identify the geospatial association between the presence of tobacco retail outlets (TRO) around schools' neighborhoods, and current use of cigarettes and e-cigarettes among adolescents in four counties in Texas. Students in grades 6, 8 and 10th were surveyed in their schools in 2014-2015. The schools' addresses was geocoded to determine the presence of at least one TRO within half a mile of the school. Two outcomes were considered: past 30-day use of (a) cigarettes and (b) e-cigarettes. Bayesian structured additive regression models and Kriging methods were used to estimate the geospatial associations between the presence of TRO and use in three counties: Dallas/Tarrant, Harris, and Travis. We observed a geospatial association between the presence of TRO around the schools and current use of cigarettes in the eastern area of Dallas County and in the southeastern area of Harris County. Also, a geospatial association between the presence of TRO around the schools and current use of e-cigarettes was observed in the entire Tarrant County and in the northeastern area of Harris County. There were geospatial associations between the presence of TRO around some schools and cigarette/e-cigarette use among students, but this association was not consistent across all the counties. More research is needed to determine why some areas are at higher risk for this association.

  2. Catalogue of gamma rays from radionuclides ordered by nuclide

    International Nuclear Information System (INIS)

    Ekstroem, L.P.; Andersson, P.; Sheppard, H.M.

    1984-01-01

    A catalogue of about 28500 gamma-ray energies from 2338 radionuclides is presented. The nuclides are listed in order of increasing (A,Z) of the daughter nuclide. In addition the gamma-ray intensity per 100 decays of the parent (if known) and the decay half-life are given. All data are from a computer processing of a recent ENSDF (Evaluated Nuclear Structure Data File) file. (authors)

  3. Temporal and spatial variations in the magnitude of completeness for homogenized moment magnitude catalogue for northeast India

    Science.gov (United States)

    Das, Ranjit; Wason, H. R.; Sharma, M. L.

    2012-02-01

    Northeast India region is one of the most seismically active areas in the world. Events data for the period 1897-2010, used in this study has been largely compiled from global ISC, NEIC and GCMT databases. Historical seismicity catalogue of Gupta et al (1986) and some events data from the bulletins of India Meteorological Department are also used. Orthogonal regression relations for conversion of body and surface wave magnitudes to M w,HRVD based on events data for the period 1978-2006 have been derived. An Orthogonal Standard Regression (OSR) relationship has also been obtained for scaling of intensity estimates to M w,NEIC using 126 global intensity events with intensity VI or greater during the period 1975-2010. Magnitude of completeness and Gutenberg-Richter (GR) recurrence parameter values have been determined for the declustered homogenized catalogue pertaining to four different time periods namely, 1897-1963, 1964-1990, 1964-2000 and 1964-2010. The M c and ` b' values are observed to decrease and increase, respectively, with addition of newer data with time. The study region has been subdivided into nine seismogenic zones keeping in view the spatial variations in earthquake occurrence and prevalent tectonics. M c, ` b' and ` a' values have been estimated with respect to each zone, and the variations in the values of these parameters have been analysed.

  4. Magnetic compatibility of standard components for electrical installations: Computation of the background field and consequences on the design of the electrical distribution boards and control boards for the ITER Tokamak building

    International Nuclear Information System (INIS)

    Benfatto, I.; Bettini, P.; Cavinato, M.; Lorenzi, A. De; Hourtoule, J.; Serra, E.

    2005-01-01

    Inside the proposed Tokamak building, the ITER poloidal field magnet system would produce a stray magnetic field up to 70 mT. This is a very unusual environmental condition for electrical installation equipment and limited information is available on the magnetic compatibility of standard components for electrical distribution boards and control boards. Because this information is a necessary input for the design of the electrical installation inside the proposed ITER Tokamak building specific investigations have been carried out by the ITER European Participant Team. The paper reports on the computation of the background magnetic field map inside the ITER Tokamak building and the consequences on the design of the electrical installations of this building. The effects of the steel inside the building structure and the feasibility of magnetic shields for electrical distribution boards and control boards are also reported in the paper. The results of the test campaigns on the magnetic field compatibility of standard components for electrical distribution boards and control boards are reported in companion papers published in these proceedings

  5. Quantifying uncertainty in NDSHA estimates due to earthquake catalogue

    Science.gov (United States)

    Magrin, Andrea; Peresan, Antonella; Vaccari, Franco; Panza, Giuliano

    2014-05-01

    The procedure for the neo-deterministic seismic zoning, NDSHA, is based on the calculation of synthetic seismograms by the modal summation technique. This approach makes use of information about the space distribution of large magnitude earthquakes, which can be defined based on seismic history and seismotectonics, as well as incorporating information from a wide set of geological and geophysical data (e.g., morphostructural features and ongoing deformation processes identified by earth observations). Hence the method does not make use of attenuation models (GMPE), which may be unable to account for the complexity of the product between seismic source tensor and medium Green function and are often poorly constrained by the available observations. NDSHA defines the hazard from the envelope of the values of ground motion parameters determined considering a wide set of scenario earthquakes; accordingly, the simplest outcome of this method is a map where the maximum of a given seismic parameter is associated to each site. In NDSHA uncertainties are not statistically treated as in PSHA, where aleatory uncertainty is traditionally handled with probability density functions (e.g., for magnitude and distance random variables) and epistemic uncertainty is considered by applying logic trees that allow the use of alternative models and alternative parameter values of each model, but the treatment of uncertainties is performed by sensitivity analyses for key modelling parameters. To fix the uncertainty related to a particular input parameter is an important component of the procedure. The input parameters must account for the uncertainty in the prediction of fault radiation and in the use of Green functions for a given medium. A key parameter is the magnitude of sources used in the simulation that is based on catalogue informations, seismogenic zones and seismogenic nodes. Because the largest part of the existing catalogues is based on macroseismic intensity, a rough estimate

  6. Geospatial-temporal semantic graph representations of trajectories from remote sensing and geolocation data

    Science.gov (United States)

    Perkins, David Nikolaus; Brost, Randolph; Ray, Lawrence P.

    2017-08-08

    Various technologies for facilitating analysis of large remote sensing and geolocation datasets to identify features of interest are described herein. A search query can be submitted to a computing system that executes searches over a geospatial temporal semantic (GTS) graph to identify features of interest. The GTS graph comprises nodes corresponding to objects described in the remote sensing and geolocation datasets, and edges that indicate geospatial or temporal relationships between pairs of nodes in the nodes. Trajectory information is encoded in the GTS graph by the inclusion of movable nodes to facilitate searches for features of interest in the datasets relative to moving objects such as vehicles.

  7. Preparing Preservice Teachers to Incorporate Geospatial Technologies in Geography Teaching

    Science.gov (United States)

    Harte, Wendy

    2017-01-01

    This study evaluated the efficacy of geospatial technology (GT) learning experiences in two geography curriculum courses to determine their effectiveness for developing preservice teacher confidence and preparing preservice teachers to incorporate GT in their teaching practices. Surveys were used to collect data from preservice teachers at three…

  8. Geospatial Analysis of Renewable Energy Technical Potential on Tribal Lands

    Energy Technology Data Exchange (ETDEWEB)

    Doris, E.; Lopez, A.; Beckley, D.

    2013-02-01

    This technical report uses an established geospatial methodology to estimate the technical potential for renewable energy on tribal lands for the purpose of allowing Tribes to prioritize the development of renewable energy resources either for community scale on-tribal land use or for revenue generating electricity sales.

  9. Large Scale Analysis of Geospatial Data with Dask and XArray

    Science.gov (United States)

    Zender, C. S.; Hamman, J.; Abernathey, R.; Evans, K. J.; Rocklin, M.; Zender, C. S.; Rocklin, M.

    2017-12-01

    The analysis of geospatial data with high level languages has acceleratedinnovation and the impact of existing data resources. However, as datasetsgrow beyond single-machine memory, data structures within these high levellanguages can become a bottleneck. New libraries like Dask and XArray resolve some of these scalability issues,providing interactive workflows that are both familiar tohigh-level-language researchers while also scaling out to much largerdatasets. This broadens the access of researchers to larger datasets on highperformance computers and, through interactive development, reducestime-to-insight when compared to traditional parallel programming techniques(MPI). This talk describes Dask, a distributed dynamic task scheduler, Dask.array, amulti-dimensional array that copies the popular NumPy interface, and XArray,a library that wraps NumPy/Dask.array with labeled and indexes axes,implementing the CF conventions. We discuss both the basic design of theselibraries and how they change interactive analysis of geospatial data, and alsorecent benefits and challenges of distributed computing on clusters ofmachines.

  10. Catalogue of Videorecordings and Films, Kindergarten to Grade 6, 1993.

    Science.gov (United States)

    Manitoba Dept. of Education, Winnipeg. Instructional Resources Branch.

    This catalogue lists and indexes 2,233 videorecordings, 16mm film, and videodisc titles held by the Library, Manitoba Education and Training for borrowing; some are also available for dubbing. The catalog indexes materials intended for children in kindergarten through grade 6, and is divided into three parts: an annotated title and series index, a…

  11. Initial PDS4 Support for the Geospatial Data Abstraction Library (GDAL)

    Science.gov (United States)

    Hare, T. M.; Gaddis, L. R.

    2018-04-01

    We introduce initial support for PDS4 within the Geospatial Data Abstraction Library (GDAL). Both highlights and limitations are presented, as well as a short discussion on methods for supporting a GDAL-based workflow for PDS4 conversions.

  12. JPEG2000-Compatible Scalable Scheme for Wavelet-Based Video Coding

    Directory of Open Access Journals (Sweden)

    Thomas André

    2007-03-01

    Full Text Available We present a simple yet efficient scalable scheme for wavelet-based video coders, able to provide on-demand spatial, temporal, and SNR scalability, and fully compatible with the still-image coding standard JPEG2000. Whereas hybrid video coders must undergo significant changes in order to support scalability, our coder only requires a specific wavelet filter for temporal analysis, as well as an adapted bit allocation procedure based on models of rate-distortion curves. Our study shows that scalably encoded sequences have the same or almost the same quality than nonscalably encoded ones, without a significant increase in complexity. A full compatibility with Motion JPEG2000, which tends to be a serious candidate for the compression of high-definition video sequences, is ensured.

  13. JPEG2000-Compatible Scalable Scheme for Wavelet-Based Video Coding

    Directory of Open Access Journals (Sweden)

    André Thomas

    2007-01-01

    Full Text Available We present a simple yet efficient scalable scheme for wavelet-based video coders, able to provide on-demand spatial, temporal, and SNR scalability, and fully compatible with the still-image coding standard JPEG2000. Whereas hybrid video coders must undergo significant changes in order to support scalability, our coder only requires a specific wavelet filter for temporal analysis, as well as an adapted bit allocation procedure based on models of rate-distortion curves. Our study shows that scalably encoded sequences have the same or almost the same quality than nonscalably encoded ones, without a significant increase in complexity. A full compatibility with Motion JPEG2000, which tends to be a serious candidate for the compression of high-definition video sequences, is ensured.

  14. Arab Libraries’ Web-based OPACs: An evaluative study in the light of IFLA’s Guidelines For Online Public Access Catalogue (OPAC Displays

    Directory of Open Access Journals (Sweden)

    Sherif Kamel Shaheen

    2005-03-01

    Full Text Available The research aims at evaluating Arabic Libraries’ Web-based Catalogues in the light of Principles and Recommendations published in: IFLA’s Guidelines For OPAC Displays (September 30, 2003 Draft For Worldwide Review. The total No. Of Recommendations reached” 38 “were categorized under three main titles, as follows: User Needs (12 recommendations, Content and arrangement Principle (25 recommendations, Standardization Principle (one recommendation However that number increased to reach 88 elements when formulated as evaluative criteria and included in the study’s checklist.

  15. The library catalogue as a retrieval tool in academic libraries: a case ...

    African Journals Online (AJOL)

    The library catalogue as a retrieval tool in academic libraries: a case study of federal university of technology Owerri library. ... Information Technologist (The) ... be emphasized and made compulsory for a ll library users and strengthening the ...

  16. Irpinia 1980, Southern Latium and Abruzzi National Park 1984, Zafferana Etnea (Sicily) 1984, Mexico City 1985 earthquakes: Catalogues of damage

    International Nuclear Information System (INIS)

    Sofo, G.

    1987-01-01

    The iconographic and descriptive documentation is collected in catalogues which are compiled in the most objective form in order to permit: (1) the determination of the behaviour of typologies and structural details; (2) the assessment of their relations with the aspects of buildings as a whole; (3) the comparison with analogue or sharply different typologies. The aim of catalogues is to establish an irreplaceble experimental basis for studies devoted to: - verify deficiencies or properties of design and building techniques; - verify possible insufficiency of antiseismic regulations; - establish univoque correlations between structural effects and macroseismic intensity. Catalogues give also reliable information for building seismic instrumentation programs, e.g. the one that ENEA is developing in cooperation with the Ministry of Works. Catalogues have been framed into a necessary compromise between representative completeness and the exigencies of survey over a sufficiently large number of buildings. (orig./HP)

  17. Development, management and benefit from Internet-based geospatial data sources through knowledge management for GIS-based regional geography applications

    International Nuclear Information System (INIS)

    Thunemann, H.G.

    2009-01-01

    The provision of data and information on the Internet is growing daily. For geoscientific applications, especially using geographic information systems (GIS), changing geospacial data are often needed, and thus possibly different data sources. Geospatial data should be easily available. As an increasingly important medium for exchange of geospatial data is the internet. The problem of finding appropriate datasources on the Internet remains to the user. The Internet as a technical basis, which was designed as a tool for information exchange, has changed the practice of dealing with knowledge and information on fundamental and not previously foreseeable manner. From the many individual acts social consequences result, concerning the production and disposal of knowledge. These determine the development of different solutions significantly, which also includes the production, deployment and use of geospatial data, with all its strengths and problems. Various solutions to the provision of geospatial data are available on the Internet, the targeted searching of this geodata sources on the Internet remains a shortcoming. The options of knowledge management, among other solutions, could be a possibility to ease the compilation, storage, connection, popularization and ultimately the application of geodata sources on the Internet. Communication, as a central element of the use of knowledge management, should be used in the form of a communication platform. The present study describes the variety of deployment options of geospatial data and the problems of finding data sources on the Internet. Potential hazards of geospatial data provision (also) via the Internet as well as an option to manage, update and use them for various applications on the Internet are are pointed out. (author) [de

  18. Rate-Compatible LDPC Codes with Linear Minimum Distance

    Science.gov (United States)

    Divsalar, Dariush; Jones, Christopher; Dolinar, Samuel

    2009-01-01

    A recently developed method of constructing protograph-based low-density parity-check (LDPC) codes provides for low iterative decoding thresholds and minimum distances proportional to block sizes, and can be used for various code rates. A code constructed by this method can have either fixed input block size or fixed output block size and, in either case, provides rate compatibility. The method comprises two submethods: one for fixed input block size and one for fixed output block size. The first mentioned submethod is useful for applications in which there are requirements for rate-compatible codes that have fixed input block sizes. These are codes in which only the numbers of parity bits are allowed to vary. The fixed-output-blocksize submethod is useful for applications in which framing constraints are imposed on the physical layers of affected communication systems. An example of such a system is one that conforms to one of many new wireless-communication standards that involve the use of orthogonal frequency-division modulation

  19. User experience and compatibility in documentation standards. Summary

    International Nuclear Information System (INIS)

    Maskewitz, B.F.

    1982-01-01

    Existing guidelines for documentation of scientific computer programs or data libraries are reviewed, and the essential elements for facilitating exchange of the software are outlined. Selected case studies are made in which accepted standards were followed from the programming stage through documentation, and an analysis of user experience

  20. Geospatial Data Repository. Sharing Data Across the Organization and Beyond

    National Research Council Canada - National Science Library

    Ruiz, Marilyn

    2001-01-01

    .... This short Technical Note discusses a five-part approach to creating a data repository that addresses the problems of the historical organizational framework for geospatial data. Fort Hood, Texas was the site used to develop the prototype. A report documenting the complete study will be available in late Spring 2001.