WorldWideScience

Sample records for multiple data-stream simd

  1. Characterization of robotics parallel algorithms and mapping onto a reconfigurable SIMD machine

    Science.gov (United States)

    Lee, C. S. G.; Lin, C. T.

    1989-01-01

    The kinematics, dynamics, Jacobian, and their corresponding inverse computations are six essential problems in the control of robot manipulators. Efficient parallel algorithms for these computations are discussed and analyzed. Their characteristics are identified and a scheme on the mapping of these algorithms to a reconfigurable parallel architecture is presented. Based on the characteristics including type of parallelism, degree of parallelism, uniformity of the operations, fundamental operations, data dependencies, and communication requirement, it is shown that most of the algorithms for robotic computations possess highly regular properties and some common structures, especially the linear recursive structure. Moreover, they are well-suited to be implemented on a single-instruction-stream multiple-data-stream (SIMD) computer with reconfigurable interconnection network. The model of a reconfigurable dual network SIMD machine with internal direct feedback is introduced. A systematic procedure internal direct feedback is introduced. A systematic procedure to map these computations to the proposed machine is presented. A new scheduling problem for SIMD machines is investigated and a heuristic algorithm, called neighborhood scheduling, that reorders the processing sequence of subtasks to reduce the communication time is described. Mapping results of a benchmark algorithm are illustrated and discussed.

  2. Streaming for Functional Data-Parallel Languages

    DEFF Research Database (Denmark)

    Madsen, Frederik Meisner

    In this thesis, we investigate streaming as a general solution to the space inefficiency commonly found in functional data-parallel programming languages. The data-parallel paradigm maps well to parallel SIMD-style hardware. However, the traditional fully materializing execution strategy...... by extending two existing data-parallel languages: NESL and Accelerate. In the extensions we map bulk operations to data-parallel streams that can evaluate fully sequential, fully parallel or anything in between. By a dataflow, piecewise parallel execution strategy, the runtime system can adjust to any target...... flattening necessitates all sub-computations to materialize at the same time. For example, naive n by n matrix multiplication requires n^3 space in NESL because the algorithm contains n^3 independent scalar multiplications. For large values of n, this is completely unacceptable. We address the problem...

  3. Single-instruction multiple-data execution

    CERN Document Server

    Hughes, Christopher J

    2015-01-01

    Having hit power limitations to even more aggressive out-of-order execution in processor cores, many architects in the past decade have turned to single-instruction-multiple-data (SIMD) execution to increase single-threaded performance. SIMD execution, or having a single instruction drive execution of an identical operation on multiple data items, was already well established as a technique to efficiently exploit data parallelism. Furthermore, support for it was already included in many commodity processors. However, in the past decade, SIMD execution has seen a dramatic increase in the set of

  4. Reduction operator for wide-SIMDs reconsidered

    NARCIS (Netherlands)

    Waeijen, L.J.W.; She, D.; Corporaal, H.; He, Y.

    2014-01-01

    It has been shown that wide Single Instruction Multiple Data architectures (wide-SIMDs) can achieve high energy efficiency, especially in domains such as image and vision processing. In these and various other application domains, reduction is a frequently encountered operation, where multiple input

  5. Optimized scalar promotion with load and splat SIMD instructions

    Science.gov (United States)

    Eichenberger, Alexander E; Gschwind, Michael K; Gunnels, John A

    2013-10-29

    Mechanisms for optimizing scalar code executed on a single instruction multiple data (SIMD) engine are provided. Placement of vector operation-splat operations may be determined based on an identification of scalar and SIMD operations in an original code representation. The original code representation may be modified to insert the vector operation-splat operations based on the determined placement of vector operation-splat operations to generate a first modified code representation. Placement of separate splat operations may be determined based on identification of scalar and SIMD operations in the first modified code representation. The first modified code representation may be modified to insert or delete separate splat operations based on the determined placement of the separate splat operations to generate a second modified code representation. SIMD code may be output based on the second modified code representation for execution by the SIMD engine.

  6. Automatic SIMD vectorization of SSA-based control flow graphs

    CERN Document Server

    Karrenberg, Ralf

    2015-01-01

    Ralf Karrenberg presents Whole-Function Vectorization (WFV), an approach that allows a compiler to automatically create code that exploits data-parallelism using SIMD instructions. Data-parallel applications such as particle simulations, stock option price estimation or video decoding require the same computations to be performed on huge amounts of data. Without WFV, one processor core executes a single instance of a data-parallel function. WFV transforms the function to execute multiple instances at once using SIMD instructions. The author describes an advanced WFV algorithm that includes a v

  7. Monte Carlo simulations on SIMD computer architectures

    International Nuclear Information System (INIS)

    Burmester, C.P.; Gronsky, R.; Wille, L.T.

    1992-01-01

    In this paper algorithmic considerations regarding the implementation of various materials science applications of the Monte Carlo technique to single instruction multiple data (SIMD) computer architectures are presented. In particular, implementation of the Ising model with nearest, next nearest, and long range screened Coulomb interactions on the SIMD architecture MasPar MP-1 (DEC mpp-12000) series of massively parallel computers is demonstrated. Methods of code development which optimize processor array use and minimize inter-processor communication are presented including lattice partitioning and the use of processor array spanning tree structures for data reduction. Both geometric and algorithmic parallel approaches are utilized. Benchmarks in terms of Monte Carl updates per second for the MasPar architecture are presented and compared to values reported in the literature from comparable studies on other architectures

  8. A flexible algorithm for calculating pair interactions on SIMD architectures

    Science.gov (United States)

    Páll, Szilárd; Hess, Berk

    2013-12-01

    Calculating interactions or correlations between pairs of particles is typically the most time-consuming task in particle simulation or correlation analysis. Straightforward implementations using a double loop over particle pairs have traditionally worked well, especially since compilers usually do a good job of unrolling the inner loop. In order to reach high performance on modern CPU and accelerator architectures, single-instruction multiple-data (SIMD) parallelization has become essential. Avoiding memory bottlenecks is also increasingly important and requires reducing the ratio of memory to arithmetic operations. Moreover, when pairs only interact within a certain cut-off distance, good SIMD utilization can only be achieved by reordering input and output data, which quickly becomes a limiting factor. Here we present an algorithm for SIMD parallelization based on grouping a fixed number of particles, e.g. 2, 4, or 8, into spatial clusters. Calculating all interactions between particles in a pair of such clusters improves data reuse compared to the traditional scheme and results in a more efficient SIMD parallelization. Adjusting the cluster size allows the algorithm to map to SIMD units of various widths. This flexibility not only enables fast and efficient implementation on current CPUs and accelerator architectures like GPUs or Intel MIC, but it also makes the algorithm future-proof. We present the algorithm with an application to molecular dynamics simulations, where we can also make use of the effective buffering the method introduces.

  9. Effective SIMD Vectorization for Intel Xeon Phi Coprocessors

    Directory of Open Access Journals (Sweden)

    Xinmin Tian

    2015-01-01

    Full Text Available Efficiently exploiting SIMD vector units is one of the most important aspects in achieving high performance of the application code running on Intel Xeon Phi coprocessors. In this paper, we present several effective SIMD vectorization techniques such as less-than-full-vector loop vectorization, Intel MIC specific alignment optimization, and small matrix transpose/multiplication 2D vectorization implemented in the Intel C/C++ and Fortran production compilers for Intel Xeon Phi coprocessors. A set of workloads from several application domains is employed to conduct the performance study of our SIMD vectorization techniques. The performance results show that we achieved up to 12.5x performance gain on the Intel Xeon Phi coprocessor. We also demonstrate a 2000x performance speedup from the seamless integration of SIMD vectorization and parallelization.

  10. Parallel computing of physical maps--a comparative study in SIMD and MIMD parallelism.

    Science.gov (United States)

    Bhandarkar, S M; Chirravuri, S; Arnold, J

    1996-01-01

    Ordering clones from a genomic library into physical maps of whole chromosomes presents a central computational problem in genetics. Chromosome reconstruction via clone ordering is usually isomorphic to the NP-complete Optimal Linear Arrangement problem. Parallel SIMD and MIMD algorithms for simulated annealing based on Markov chain distribution are proposed and applied to the problem of chromosome reconstruction via clone ordering. Perturbation methods and problem-specific annealing heuristics are proposed and described. The SIMD algorithms are implemented on a 2048 processor MasPar MP-2 system which is an SIMD 2-D toroidal mesh architecture whereas the MIMD algorithms are implemented on an 8 processor Intel iPSC/860 which is an MIMD hypercube architecture. A comparative analysis of the various SIMD and MIMD algorithms is presented in which the convergence, speedup, and scalability characteristics of the various algorithms are analyzed and discussed. On a fine-grained, massively parallel SIMD architecture with a low synchronization overhead such as the MasPar MP-2, a parallel simulated annealing algorithm based on multiple periodically interacting searches performs the best. For a coarse-grained MIMD architecture with high synchronization overhead such as the Intel iPSC/860, a parallel simulated annealing algorithm based on multiple independent searches yields the best results. In either case, distribution of clonal data across multiple processors is shown to exacerbate the tendency of the parallel simulated annealing algorithm to get trapped in a local optimum.

  11. Effective SIMD Vectorization for Intel Xeon Phi Coprocessors

    OpenAIRE

    Tian, Xinmin; Saito, Hideki; Preis, Serguei V.; Garcia, Eric N.; Kozhukhov, Sergey S.; Masten, Matt; Cherkasov, Aleksei G.; Panchenko, Nikolay

    2015-01-01

    Efficiently exploiting SIMD vector units is one of the most important aspects in achieving high performance of the application code running on Intel Xeon Phi coprocessors. In this paper, we present several effective SIMD vectorization techniques such as less-than-full-vector loop vectorization, Intel MIC specific alignment optimization, and small matrix transpose/multiplication 2D vectorization implemented in the Intel C/C++ and Fortran production compilers for Intel Xeon Phi coprocessors. A ...

  12. High performance multiple stream data transfer

    International Nuclear Information System (INIS)

    Rademakers, F.; Saiz, P.

    2001-01-01

    The ALICE detector at LHC (CERN), will record raw data at a rate of 1.2 Gigabytes per second. Trying to analyse all this data at CERN will not be feasible. As originally proposed by the MONARC project, data collected at CERN will be transferred to remote centres to use their computing infrastructure. The remote centres will reconstruct and analyse the events, and make available the results. Therefore high-rate data transfer between computing centres (Tiers) will become of paramount importance. The authors will present several tests that have been made between CERN and remote centres in Padova (Italy), Torino (Italy), Catania (Italy), Lyon (France), Ohio (United States), Warsaw (Poland) and Calcutta (India). These tests consisted, in a first stage, of sending raw data from CERN to the remote centres and back, using a ftp method that allows connections of several streams at the same time. Thanks to these multiple streams, it is possible to increase the rate at which the data is transferred. While several 'multiple stream ftp solutions' already exist, the authors' method is based on a parallel socket implementation which allows, besides files, also objects (or any large message) to be send in parallel. A prototype will be presented able to manage different transfers. This is the first step of a system to be implemented that will be able to take care of the connections with the remote centres to exchange data and monitor the status of the transfer

  13. A language for data-parallel and task parallel programming dedicated to multi-SIMD computers. Contributions to hydrodynamic simulation with lattice gases

    International Nuclear Information System (INIS)

    Pic, Marc Michel

    1995-01-01

    Parallel programming covers task-parallelism and data-parallelism. Many problems need both parallelisms. Multi-SIMD computers allow hierarchical approach of these parallelisms. The T++ language, based on C++, is dedicated to exploit Multi-SIMD computers using a programming paradigm which is an extension of array-programming to tasks managing. Our language introduced array of independent tasks to achieve separately (MIMD), on subsets of processors of identical behaviour (SIMD), in order to translate the hierarchical inclusion of data-parallelism in task-parallelism. To manipulate in a symmetrical way tasks and data we propose meta-operations which have the same behaviour on tasks arrays and on data arrays. We explain how to implement this language on our parallel computer SYMPHONIE in order to profit by the locally-shared memory, by the hardware virtualization, and by the multiplicity of communications networks. We analyse simultaneously a typical application of such architecture. Finite elements scheme for Fluid mechanic needs powerful parallel computers and requires large floating points abilities. Lattice gases is an alternative to such simulations. Boolean lattice bases are simple, stable, modular, need to floating point computation, but include numerical noise. Boltzmann lattice gases present large precision of computation, but needs floating points and are only locally stable. We propose a new scheme, called multi-bit, who keeps the advantages of each boolean model to which it is applied, with large numerical precision and reduced noise. Experiments on viscosity, physical behaviour, noise reduction and spurious invariants are shown and implementation techniques for parallel Multi-SIMD computers detailed. (author) [fr

  14. Rapid prototyping and evaluation of programmable SIMD SDR processors in LISA

    Science.gov (United States)

    Chen, Ting; Liu, Hengzhu; Zhang, Botao; Liu, Dongpei

    2013-03-01

    With the development of international wireless communication standards, there is an increase in computational requirement for baseband signal processors. Time-to-market pressure makes it impossible to completely redesign new processors for the evolving standards. Due to its high flexibility and low power, software defined radio (SDR) digital signal processors have been proposed as promising technology to replace traditional ASIC and FPGA fashions. In addition, there are large numbers of parallel data processed in computation-intensive functions, which fosters the development of single instruction multiple data (SIMD) architecture in SDR platform. So a new way must be found to prototype the SDR processors efficiently. In this paper we present a bit-and-cycle accurate model of programmable SIMD SDR processors in a machine description language LISA. LISA is a language for instruction set architecture which can gain rapid model at architectural level. In order to evaluate the availability of our proposed processor, three common baseband functions, FFT, FIR digital filter and matrix multiplication have been mapped on the SDR platform. Analytical results showed that the SDR processor achieved the maximum of 47.1% performance boost relative to the opponent processor.

  15. Architecture Without Explicit Locks for Logic Simulation on SIMD Machines

    OpenAIRE

    Cockshott, W. Paul; Chimeh, Mozhgan Kabiri

    2016-01-01

    The presentation describes an architecture for logic simulation that takes advantages of the features of multi-core SIMD architectures. It uses neither explicit locks nor queues, using instead oblivious simulation. Data structures are targeted to efficient SIMD and multi-core cache operation. We demonstrate high levels of parallelisation on Xeon Phi and AMD multi-core machines. Performance on a Xeon Phi is comparable to or better than on a 1000 core Blue Gene machine.

  16. Floating point only SIMD instruction set architecture including compare, select, Boolean, and alignment operations

    Science.gov (United States)

    Gschwind, Michael K [Chappaqua, NY

    2011-03-01

    Mechanisms for implementing a floating point only single instruction multiple data instruction set architecture are provided. A processor is provided that comprises an issue unit, an execution unit coupled to the issue unit, and a vector register file coupled to the execution unit. The execution unit has logic that implements a floating point (FP) only single instruction multiple data (SIMD) instruction set architecture (ISA). The floating point vector registers of the vector register file store both scalar and floating point values as vectors having a plurality of vector elements. The processor may be part of a data processing system.

  17. Coupling SIMD and SIMT architectures to boost performance of a phylogeny-aware alignment kernel

    Directory of Open Access Journals (Sweden)

    Alachiotis Nikolaos

    2012-08-01

    Full Text Available Abstract Background Aligning short DNA reads to a reference sequence alignment is a prerequisite for detecting their biological origin and analyzing them in a phylogenetic context. With the PaPaRa tool we introduced a dedicated dynamic programming algorithm for simultaneously aligning short reads to reference alignments and corresponding evolutionary reference trees. The algorithm aligns short reads to phylogenetic profiles that correspond to the branches of such a reference tree. The algorithm needs to perform an immense number of pairwise alignments. Therefore, we explore vector intrinsics and GPUs to accelerate the PaPaRa alignment kernel. Results We optimized and parallelized PaPaRa on CPUs and GPUs. Via SSE 4.1 SIMD (Single Instruction, Multiple Data intrinsics for x86 SIMD architectures and multi-threading, we obtained a 9-fold acceleration on a single core as well as linear speedups with respect to the number of cores. The peak CPU performance amounts to 18.1 GCUPS (Giga Cell Updates per Second using all four physical cores on an Intel i7 2600 CPU running at 3.4 GHz. The average CPU performance (averaged over all test runs is 12.33 GCUPS. We also used OpenCL to execute PaPaRa on a GPU SIMT (Single Instruction, Multiple Threads architecture. A NVIDIA GeForce 560 GPU delivered peak and average performance of 22.1 and 18.4 GCUPS respectively. Finally, we combined the SIMD and SIMT implementations into a hybrid CPU-GPU system that achieved an accumulated peak performance of 33.8 GCUPS. Conclusions This accelerated version of PaPaRa (available at http://www.exelixis-lab.org/software.html provides a significant performance improvement that allows for analyzing larger datasets in less time. We observe that state-of-the-art SIMD and SIMT architectures deliver comparable performance for this dynamic programming kernel when the “competing programmer approach” is deployed. Finally, we show that overall performance can be substantially increased

  18. Generating and executing programs for a floating point single instruction multiple data instruction set architecture

    Science.gov (United States)

    Gschwind, Michael K

    2013-04-16

    Mechanisms for generating and executing programs for a floating point (FP) only single instruction multiple data (SIMD) instruction set architecture (ISA) are provided. A computer program product comprising a computer recordable medium having a computer readable program recorded thereon is provided. The computer readable program, when executed on a computing device, causes the computing device to receive one or more instructions and execute the one or more instructions using logic in an execution unit of the computing device. The logic implements a floating point (FP) only single instruction multiple data (SIMD) instruction set architecture (ISA), based on data stored in a vector register file of the computing device. The vector register file is configured to store both scalar and floating point values as vectors having a plurality of vector elements.

  19. Exploring query execution strategies for JIT vectorization and SIMD

    NARCIS (Netherlands)

    T.K. Gubner (Tim); P.A. Boncz (Peter)

    2017-01-01

    textabstractThis paper partially explores the design space for efficient query processors on future hardware that is rich in SIMD capabilities. It departs from two well-known approaches: (1) interpreted block-at-a-time execution (a.k.a. "vectorization") and (2) "data-centric" JIT compilation, as in

  20. Digital dashboard design using multiple data streams for disease surveillance with influenza surveillance as an example.

    Science.gov (United States)

    Cheng, Calvin K Y; Ip, Dennis K M; Cowling, Benjamin J; Ho, Lai Ming; Leung, Gabriel M; Lau, Eric H Y

    2011-10-14

    Great strides have been made exploring and exploiting new and different sources of disease surveillance data and developing robust statistical methods for analyzing the collected data. However, there has been less research in the area of dissemination. Proper dissemination of surveillance data can facilitate the end user's taking of appropriate actions, thus maximizing the utility of effort taken from upstream of the surveillance-to-action loop. The aims of the study were to develop a generic framework for a digital dashboard incorporating features of efficient dashboard design and to demonstrate this framework by specific application to influenza surveillance in Hong Kong. Based on the merits of the national websites and principles of efficient dashboard design, we designed an automated influenza surveillance digital dashboard as a demonstration of efficient dissemination of surveillance data. We developed the system to synthesize and display multiple sources of influenza surveillance data streams in the dashboard. Different algorithms can be implemented in the dashboard for incorporating all surveillance data streams to describe the overall influenza activity. We designed and implemented an influenza surveillance dashboard that utilized self-explanatory figures to display multiple surveillance data streams in panels. Indicators for individual data streams as well as for overall influenza activity were summarized in the main page, which can be read at a glance. Data retrieval function was also incorporated to allow data sharing in standard format. The influenza surveillance dashboard serves as a template to illustrate the efficient synthesization and dissemination of multiple-source surveillance data, which may also be applied to other diseases. Surveillance data from multiple sources can be disseminated efficiently using a dashboard design that facilitates the translation of surveillance information to public health actions.

  1. Streaming nested data parallelism on multicores

    DEFF Research Database (Denmark)

    Madsen, Frederik Meisner; Filinski, Andrzej

    2016-01-01

    The paradigm of nested data parallelism (NDP) allows a variety of semi-regular computation tasks to be mapped onto SIMD-style hardware, including GPUs and vector units. However, some care is needed to keep down space consumption in situations where the available parallelism may vastly exceed...

  2. MoviCompile : An LLVM based compiler for heterogeneous SIMD code generation

    NARCIS (Netherlands)

    Diken, E.; Jordans, R.; O'Riordan, M.

    2015-01-01

    Numerous applications in communication and multimedia domains show significant data-level parallelism (DLP). The amount of DLP varies between applications in the same domain or even within a single application. Most architectures support a single vector-, SIMD-width which may not be optimal. This

  3. OpenCL code generation for low energy wide SIMD architectures with explicit datapath.

    NARCIS (Netherlands)

    She, D.; He, Y.; Waeijen, L.J.W.; Corporaal, H.; Jeschke, H.; Silvén, O.

    2013-01-01

    Energy efficiency is one of the most important aspects in designing embedded processors. The use of a wide SIMD processor architecture is a promising approach to build energy-efficient high performance embedded processors. In this paper, we propose a configurable wide SIMD architecture that utilizes

  4. Filtering Redundant Data from RFID Data Streams

    Directory of Open Access Journals (Sweden)

    Hazalila Kamaludin

    2016-01-01

    Full Text Available Radio Frequency Identification (RFID enabled systems are evolving in many applications that need to know the physical location of objects such as supply chain management. Naturally, RFID systems create large volumes of duplicate data. As the duplicate data wastes communication, processing, and storage resources as well as delaying decision-making, filtering duplicate data from RFID data stream is an important and challenging problem. Existing Bloom Filter-based approaches for filtering duplicate RFID data streams are complex and slow as they use multiple hash functions. In this paper, we propose an approach for filtering duplicate data from RFID data streams. The proposed approach is based on modified Bloom Filter and uses only a single hash function. We performed extensive empirical study of the proposed approach and compared it against the Bloom Filter, d-Left Time Bloom Filter, and the Count Bloom Filter approaches. The results show that the proposed approach outperforms the baseline approaches in terms of false positive rate, execution time, and true positive rate.

  5. Search of molecular ground state via genetic algorithm: Implementation on a hybrid SIMD-MIMD platform

    International Nuclear Information System (INIS)

    Pucello, N.; D'Agostino, G.; Pisacane, F.

    1997-01-01

    A genetic algorithm for the optimization of the ground-state structure of a metallic cluster has been developed and ported on a SIMD-MIMD parallel platform. The SIMD part of the parallel platform is represented by a Quadrics/APE100 consisting of 512 floating point units, while the MIMD part is formed by a cluster of workstations. The proposed algorithm is composed by a part where the genetic operators are applied to the elements of the population and a part which performs a further local relaxation and the fitness calculation via Molecular Dynamics. These parts have been implemented on the MIMD and on the SIMD part, respectively. Results have been compared to those generated by using Simulated Annealing

  6. N-body simulation for self-gravitating collisional systems with a new SIMD instruction set extension to the x86 architecture, Advanced Vector eXtensions

    Science.gov (United States)

    Tanikawa, Ataru; Yoshikawa, Kohji; Okamoto, Takashi; Nitadori, Keigo

    2012-02-01

    We present a high-performance N-body code for self-gravitating collisional systems accelerated with the aid of a new SIMD instruction set extension of the x86 architecture: Advanced Vector eXtensions (AVX), an enhanced version of the Streaming SIMD Extensions (SSE). With one processor core of Intel Core i7-2600 processor (8 MB cache and 3.40 GHz) based on Sandy Bridge micro-architecture, we implemented a fourth-order Hermite scheme with individual timestep scheme ( Makino and Aarseth, 1992), and achieved the performance of ˜20 giga floating point number operations per second (GFLOPS) for double-precision accuracy, which is two times and five times higher than that of the previously developed code implemented with the SSE instructions ( Nitadori et al., 2006b), and that of a code implemented without any explicit use of SIMD instructions with the same processor core, respectively. We have parallelized the code by using so-called NINJA scheme ( Nitadori et al., 2006a), and achieved ˜90 GFLOPS for a system containing more than N = 8192 particles with 8 MPI processes on four cores. We expect to achieve about 10 tera FLOPS (TFLOPS) for a self-gravitating collisional system with N ˜ 10 5 on massively parallel systems with at most 800 cores with Sandy Bridge micro-architecture. This performance will be comparable to that of Graphic Processing Unit (GPU) cluster systems, such as the one with about 200 Tesla C1070 GPUs ( Spurzem et al., 2010). This paper offers an alternative to collisional N-body simulations with GRAPEs and GPUs.

  7. Vectorization with SIMD extensions speeds up reconstruction in electron tomography.

    Science.gov (United States)

    Agulleiro, J I; Garzón, E M; García, I; Fernández, J J

    2010-06-01

    Electron tomography allows structural studies of cellular structures at molecular detail. Large 3D reconstructions are needed to meet the resolution requirements. The processing time to compute these large volumes may be considerable and so, high performance computing techniques have been used traditionally. This work presents a vector approach to tomographic reconstruction that relies on the exploitation of the SIMD extensions available in modern processors in combination to other single processor optimization techniques. This approach succeeds in producing full resolution tomograms with an important reduction in processing time, as evaluated with the most common reconstruction algorithms, namely WBP and SIRT. The main advantage stems from the fact that this approach is to be run on standard computers without the need of specialized hardware, which facilitates the development, use and management of programs. Future trends in processor design open excellent opportunities for vector processing with processor's SIMD extensions in the field of 3D electron microscopy.

  8. Time-Based Data Streams: Fundamental Concepts for a Data Resource for Streams

    Energy Technology Data Exchange (ETDEWEB)

    Beth A. Plale

    2009-10-10

    Real time data, which we call data streams, are readings from instruments, environmental, bodily or building sensors that are generated at regular intervals and often, due to their volume, need to be processed in real time. Often a single pass is all that can be made on the data, and a decision to discard or keep the instance is made on the spot. Too, the stream is for all practical purposes indefinite, so decisions must be made on incomplete knowledge. This notion of data streams has a different set of issues from a file, for instance, that is byte streamed to a reader. The file is finite, so the byte stream is becomes a processing convenience more than a fundamentally different kind of data. Through the duration of the project we examined three aspects of streaming data: the first, techniques to handle streaming data in a distributed system organized as a collection of web services, the second, the notion of the dashboard and real time controllable analysis constructs in the context of the Fermi Tevatron Beam Position Monitor, and third and finally, we examined provenance collection of stream processing such as might occur as raw observational data flows from the source and undergoes correction, cleaning, and quality control. The impact of this work is severalfold. We were one of the first to advocate that streams had little value unless aggregated, and that notion is now gaining general acceptance. We were one of the first groups to grapple with the notion of provenance of stream data also.

  9. Real time image synthesis on a SIMD linear array processor: algorithms and architectures

    International Nuclear Information System (INIS)

    Letellier, Laurent

    1993-01-01

    Nowadays, image synthesis has become a widely used technique. The impressive computing power required for real time applications necessitates the use of parallel architectures. In this context, we evaluate an SIMD linear parallel architecture, SYMPATI2, dedicated to image processing. The objective of this study is to propose a cost-effective graphics accelerator relying on SYMPATI2's modular and programmable structure. The parallelization of basic image synthesis algorithms on SYMPATI2 enables us to determine its limits in this application field. These limits lead us to evaluate a new structure with a fast intercommunication network between processors, but processors have to support the message consistency, which brings about a strong decrease in performance. To solve this problem, we suggest a simple network whose access priorities are represented by tokens. The simulations of this new architecture indicate that the SIMD mode causes a drastic cut in parallelism. To cope with this drawback, we propose a context switching procedure which reduces the SIMD rigidity and increases the parallelism rate significantly. Then, the graphics accelerator we propose is compared with existing graphics workstations. This comparison indicates that our structure, which is able to accelerate both image synthesis and image processing, is competitive and well-suited for multimedia applications. (author) [fr

  10. Optimizing Cost of Continuous Overlapping Queries over Data Streams by Filter Adaption

    KAUST Repository

    Xie, Qing; Zhang, Xiangliang; Li, Zhixu; Zhou, Xiaofang

    2016-01-01

    The problem we aim to address is the optimization of cost management for executing multiple continuous queries on data streams, where each query is defined by several filters, each of which monitors certain status of the data stream. Specially

  11. Some multigrid algorithms for SIMD machines

    Energy Technology Data Exchange (ETDEWEB)

    Dendy, J.E. Jr. [Los Alamos National Lab., NM (United States)

    1996-12-31

    Previously a semicoarsening multigrid algorithm suitable for use on SIMD architectures was investigated. Through the use of new software tools, the performance of this algorithm has been considerably improved. The method has also been extended to three space dimensions. The method performs well for strongly anisotropic problems and for problems with coefficients jumping by orders of magnitude across internal interfaces. The parallel efficiency of this method is analyzed, and its actual performance on the CM-5 is compared with its performance on the CRAY-YMP. A standard coarsening multigrid algorithm is also considered, and we compare its performance on these two platforms as well.

  12. Use of multiple dispersal pathways facilitates amphibian persistence in stream networks

    Science.gov (United States)

    Campbell, Grant E.H.; Nichols, J.D.; Lowe, W.H.; Fagan, W.F.

    2010-01-01

    Although populations of amphibians are declining worldwide, there is no evidence that salamanders occupying small streams are experiencing enigmatic declines, and populations of these species seem stable. Theory predicts that dispersal through multiple pathways can stabilize populations, preventing extinction in habitat networks. However, empirical data to support this prediction are absent for most species, especially those at risk of decline. Our mark-recapture study of stream salamanders reveals both a strong upstream bias in dispersal and a surprisingly high rate of overland dispersal to adjacent headwater streams. This evidence of route-dependent variation in dispersal rates suggests a spatial mechanism for population stability in headwater-stream salamanders. Our results link the movement behavior of stream salamanders to network topology, and they underscore the importance of identifying and protecting critical dispersal pathways when addressing region-wide population declines.

  13. Use of multiple dispersal pathways facilitates amphibian persistence in stream networks.

    Science.gov (United States)

    Campbell Grant, Evan H; Nichols, James D; Lowe, Winsor H; Fagan, William F

    2010-04-13

    Although populations of amphibians are declining worldwide, there is no evidence that salamanders occupying small streams are experiencing enigmatic declines, and populations of these species seem stable. Theory predicts that dispersal through multiple pathways can stabilize populations, preventing extinction in habitat networks. However, empirical data to support this prediction are absent for most species, especially those at risk of decline. Our mark-recapture study of stream salamanders reveals both a strong upstream bias in dispersal and a surprisingly high rate of overland dispersal to adjacent headwater streams. This evidence of route-dependent variation in dispersal rates suggests a spatial mechanism for population stability in headwater-stream salamanders. Our results link the movement behavior of stream salamanders to network topology, and they underscore the importance of identifying and protecting critical dispersal pathways when addressing region-wide population declines.

  14. Flexbar 3.0 - SIMD and multicore parallelization.

    Science.gov (United States)

    Roehr, Johannes T; Dieterich, Christoph; Reinert, Knut

    2017-09-15

    High-throughput sequencing machines can process many samples in a single run. For Illumina systems, sequencing reads are barcoded with an additional DNA tag that is contained in the respective sequencing adapters. The recognition of barcode and adapter sequences is hence commonly needed for the analysis of next-generation sequencing data. Flexbar performs demultiplexing based on barcodes and adapter trimming for such data. The massive amounts of data generated on modern sequencing machines demand that this preprocessing is done as efficiently as possible. We present Flexbar 3.0, the successor of the popular program Flexbar. It employs now twofold parallelism: multi-threading and additionally SIMD vectorization. Both types of parallelism are used to speed-up the computation of pair-wise sequence alignments, which are used for the detection of barcodes and adapters. Furthermore, new features were included to cover a wide range of applications. We evaluated the performance of Flexbar based on a simulated sequencing dataset. Our program outcompetes other tools in terms of speed and is among the best tools in the presented quality benchmark. https://github.com/seqan/flexbar. johannes.roehr@fu-berlin.de or knut.reinert@fu-berlin.de. © The Author (2017). Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com

  15. Combining multiple approaches and optimized data resolution for an improved understanding of stream temperature dynamics of a forested headwater basin in the Southern Appalachians

    Science.gov (United States)

    Belica, L.; Mitasova, H.; Caldwell, P.; McCarter, J. B.; Nelson, S. A. C.

    2017-12-01

    Thermal regimes of forested headwater streams continue to be an area of active research as climatic, hydrologic, and land cover changes can influence water temperature, a key aspect of aquatic ecosystems. Widespread monitoring of stream temperatures have provided an important data source, yielding insights on the temporal and spatial patterns and the underlying processes that influence stream temperature. However, small forested streams remain challenging to model due to the high spatial and temporal variability of stream temperatures and the climatic and hydrologic conditions that drive them. Technological advances and increased computational power continue to provide new tools and measurement methods and have allowed spatially explicit analyses of dynamic natural systems at greater temporal resolutions than previously possible. With the goal of understanding how current stream temperature patterns and processes may respond to changing landcover and hydroclimatoligical conditions, we combined high-resolution, spatially explicit geospatial modeling with deterministic heat flux modeling approaches using data sources that ranged from traditional hydrological and climatological measurements to emerging remote sensing techniques. Initial analyses of stream temperature monitoring data revealed that high temporal resolution (5 minutes) and measurement resolutions (guide field data collection for further heat flux modeling. By integrating multiple approaches and optimizing data resolution for the processes being investigated, small, but ecologically significant differences in stream thermal regimes were revealed. In this case, multi-approach research contributed to the identification of the dominant mechanisms driving stream temperature in the study area and advanced our understanding of the current thermal fluxes and how they may change as environmental conditions change in the future.

  16. Parallel computing and networking; Heiretsu keisanki to network

    Energy Technology Data Exchange (ETDEWEB)

    Asakawa, E; Tsuru, T [Japan National Oil Corp., Tokyo (Japan); Matsuoka, T [Japan Petroleum Exploration Co. Ltd., Tokyo (Japan)

    1996-05-01

    This paper describes the trend of parallel computers used in geophysical exploration. Around 1993 was the early days when the parallel computers began to be used for geophysical exploration. Classification of these computers those days was mainly MIMD (multiple instruction stream, multiple data stream), SIMD (single instruction stream, multiple data stream) and the like. Parallel computers were publicized in the 1994 meeting of the Geophysical Exploration Society as a `high precision imaging technology`. Concerning the library of parallel computers, there was a shift to PVM (parallel virtual machine) in 1993 and to MPI (message passing interface) in 1995. In addition, the compiler of FORTRAN90 was released with support implemented for data parallel and vector computers. In 1993, networks used were Ethernet, FDDI, CDDI and HIPPI. In 1995, the OC-3 products under ATM began to propagate. However, ATM remains to be an interoffice high speed network because the ATM service has not spread yet for the public network. 1 ref.

  17. HydroCloud: A Web Application for Exploring Stream Gage Data

    Directory of Open Access Journals (Sweden)

    Martin C. Roberge

    2017-08-01

    Full Text Available HydroCloud (hydrocloud.org is a mobile-friendly web application for visually browsing hydrology data from multiple sources. Data providers such as the US Geological Survey (USGS and the German 'Wasserstraßen- und Schifffahrtsverwaltung des Bundes' (WSV currently serve stream discharge data from more than 10,000 stream gages around the world. HydroCloud allows users to plot these data while out in the field, while also providing contextual information such as the current NEXRAD weather imagery or descriptive information about the stream gage and its watershed. Additional features include a chat mechanism for contacting developers, and the use of local storage for saving data.   Funding Statement: This project was supported in part by a grant from the Towson University School of Emerging Technology.

  18. CAMS: OLAPing Multidimensional Data Streams Efficiently

    Science.gov (United States)

    Cuzzocrea, Alfredo

    In the context of data stream research, taming the multidimensionality of real-life data streams in order to efficiently support OLAP analysis/mining tasks is a critical challenge. Inspired by this fundamental motivation, in this paper we introduce CAMS (C ube-based A cquisition model for M ultidimensional S treams), a model for efficiently OLAPing multidimensional data streams. CAMS combines a set of data stream processing methodologies, namely (i) the OLAP dimension flattening process, which allows us to obtain dimensionality reduction of multidimensional data streams, and (ii) the OLAP stream aggregation scheme, which aggregates data stream readings according to an OLAP-hierarchy-based membership approach. We complete our analytical contribution by means of experimental assessment and analysis of both the efficiency and the scalability of OLAPing capabilities of CAMS on synthetic multidimensional data streams. Both analytical and experimental results clearly connote CAMS as an enabling component for next-generation Data Stream Management Systems.

  19. Parallel sorting algorithms

    CERN Document Server

    Akl, Selim G

    1985-01-01

    Parallel Sorting Algorithms explains how to use parallel algorithms to sort a sequence of items on a variety of parallel computers. The book reviews the sorting problem, the parallel models of computation, parallel algorithms, and the lower bounds on the parallel sorting problems. The text also presents twenty different algorithms, such as linear arrays, mesh-connected computers, cube-connected computers. Another example where algorithm can be applied is on the shared-memory SIMD (single instruction stream multiple data stream) computers in which the whole sequence to be sorted can fit in the

  20. Architecture and VHDL behavioural validation of a parallel processor dedicated to computer vision

    International Nuclear Information System (INIS)

    Collette, Thierry

    1992-01-01

    Speeding up image processing is mainly obtained using parallel computers; SIMD processors (single instruction stream, multiple data stream) have been developed, and have proven highly efficient regarding low-level image processing operations. Nevertheless, their performances drop for most intermediate of high level operations, mainly when random data reorganisations in processor memories are involved. The aim of this thesis was to extend the SIMD computer capabilities to allow it to perform more efficiently at the image processing intermediate level. The study of some representative algorithms of this class, points out the limits of this computer. Nevertheless, these limits can be erased by architectural modifications. This leads us to propose SYMPATIX, a new SIMD parallel computer. To valid its new concept, a behavioural model written in VHDL - Hardware Description Language - has been elaborated. With this model, the new computer performances have been estimated running image processing algorithm simulations. VHDL modeling approach allows to perform the system top down electronic design giving an easy coupling between system architectural modifications and their electronic cost. The obtained results show SYMPATIX to be an efficient computer for low and intermediate level image processing. It can be connected to a high level computer, opening up the development of new computer vision applications. This thesis also presents, a top down design method, based on the VHDL, intended for electronic system architects. (author) [fr

  1. The online performance estimation framework: heterogeneous ensemble learning for data streams

    NARCIS (Netherlands)

    van Rijn, J.N.; Holmes, G.; Pfahringer, B.; Vanschoren, J.

    2018-01-01

    Ensembles of classifiers are among the best performing classifiers available in many data mining applications, including the mining of data streams. Rather than training one classifier, multiple classifiers are trained, and their predictions are combined according to a given voting schedule. An

  2. What Can Hierarchies Do for Data Streams?

    DEFF Research Database (Denmark)

    Yin, Xuepeng; Pedersen, Torben Bach

    Much effort has been put into building data streams management systems for querying data streams. Here, data streams have been viewed as a flow of low-level data items, e.g., sensor readings or IP packet data. Stream query languages have mostly been SQL-based, with the STREAM and TelegraphCQ lang...

  3. Knowledge discovery from data streams

    CERN Document Server

    Gama, Joao

    2010-01-01

    Since the beginning of the Internet age and the increased use of ubiquitous computing devices, the large volume and continuous flow of distributed data have imposed new constraints on the design of learning algorithms. Exploring how to extract knowledge structures from evolving and time-changing data, Knowledge Discovery from Data Streams presents a coherent overview of state-of-the-art research in learning from data streams.The book covers the fundamentals that are imperative to understanding data streams and describes important applications, such as TCP/IP traffic, GPS data, sensor networks,

  4. Wadeable Streams Assessment Data

    Science.gov (United States)

    The Wadeable Streams Assessment (WSA) is a first-ever statistically-valid survey of the biological condition of small streams throughout the U.S. The U.S. Environmental Protection Agency (EPA) worked with the states to conduct the assessment in 2004-2005. Data for each parameter sampled in the Wadeable Streams Assessment (WSA) are available for downloading in a series of files as comma separated values (*.csv). Each *.csv data file has a companion text file (*.txt) that lists a dataset label and individual descriptions for each variable. Users should view the *.txt files first to help guide their understanding and use of the data.

  5. Data Stream Processing Study in a Multichannel Telemetry Data Registering System

    Directory of Open Access Journals (Sweden)

    I. M. Sidyakin

    2015-01-01

    Full Text Available The paper presents the results of research that is aimed to improve the reliability of transmission of telemetry information (TMI through a communication channel with noise from the object of telemeasurements to the telemetry system for collecting and processing data. It considers the case where the quality of received information changes over time, due to movement of the object relative to the receiving station, or other factors that cause changes in the characteristics of noise in the channel, up to the total loss due to some temporary sites. To improve the reliability of transmission and ensure continuous communication with the object, it is proposed to use a multi-channel system to record the TMI. This system consists of several telemetry stations, which simultaneously register data stream transmitted from the telemetry object. The multichannel system generates a single stream of TMI for the user at the output. The stream comprises the most reliable pieces of information, being received at all inputs of the system.The paper investigates the task of constructing a multi-channel registration scheme for telemetry information (TMI to provide a simultaneous reception of the telemeasurement data by multiple telemetry stations and to form a single TMI stream containing the most reliable pieces of received data on the basis of quality analysis of information being received.In a multichannel registering system of TMI there are three main factors affecting the quality of the output of a single stream of information: 1 quality of the method used for protecting against errors during transmission over the communication channel with noise; 2 efficiency of the synchronization process of telemetry frames in the received flow of information; 3 efficiency of the applied criteria to form a single output stream from multiple input streams coming from different stations in the discussed multichannel registering system of TMI.In the paper, in practical

  6. Multiple stress response of lowland stream benthic macroinvertebrates is dependent on habitat type

    DEFF Research Database (Denmark)

    Graeber, Daniel; Jensen, Tinna M.; Rasmussen, Jes

    2017-01-01

    Worldwide, lowland stream ecosystems are exposed to multiple anthropogenic stress due to the combination of water scarcity, eutrophication and fine sedimentation. The understanding of the effects of such multiple stress on stream benthic macroinvertebrates has been growing in the recent years...

  7. Introduction to stream: An Extensible Framework for Data Stream Clustering Research with R

    Directory of Open Access Journals (Sweden)

    Michael Hahsler

    2017-02-01

    Full Text Available In recent years, data streams have become an increasingly important area of research for the computer science, database and statistics communities. Data streams are ordered and potentially unbounded sequences of data points created by a typically non-stationary data generating process. Common data mining tasks associated with data streams include clustering, classification and frequent pattern mining. New algorithms for these types of data are proposed regularly and it is important to evaluate them thoroughly under standardized conditions. In this paper we introduce stream, a research tool that includes modeling and simulating data streams as well as an extensible framework for implementing, interfacing and experimenting with algorithms for various data stream mining tasks. The main advantage of stream is that it seamlessly integrates with the large existing infrastructure provided by R. In addition to data handling, plotting and easy scripting capabilities, R also provides many existing algorithms and enables users to interface code written in many programming languages popular among data mining researchers (e.g., C/C++, Java and Python. In this paper we describe the architecture of stream and focus on its use for data stream clustering research. stream was implemented with extensibility in mind and will be extended in the future to cover additional data stream mining tasks like classification and frequent pattern mining.

  8. On the Organization of Parallel Operation of Some Algorithms for Finding the Shortest Path on a Graph on a Computer System with Multiple Instruction Stream and Single Data Stream

    Directory of Open Access Journals (Sweden)

    V. E. Podol'skii

    2015-01-01

    Full Text Available The paper considers the implementing Bellman-Ford and Lee algorithms to find the shortest graph path on a computer system with multiple instruction stream and single data stream (MISD. The MISD computer is a computer that executes commands of arithmetic-logical processing (on the CPU and commands of structures processing (on the structures processor in parallel on a single data stream. Transformation of sequential programs into the MISD programs is a labor intensity process because it requires a stream of the arithmetic-logical processing to be manually separated from that of the structures processing. Algorithms based on the processing of data structures (e.g., algorithms on graphs show high performance on a MISD computer. Bellman-Ford and Lee algorithms for finding the shortest path on a graph are representatives of these algorithms. They are applied to robotics for automatic planning of the robot movement in-situ. Modification of Bellman-Ford and Lee algorithms for finding the shortest graph path in coprocessor MISD mode and the parallel MISD modification of these algorithms were first obtained in this article. Thus, this article continues a series of studies on the transformation of sequential algorithms into MISD ones (Dijkstra and Ford-Fulkerson 's algorithms and has a pronouncedly applied nature. The article also presents the analysis results of Bellman-Ford and Lee algorithms in MISD mode. The paper formulates the basic trends of a technique for parallelization of algorithms into arithmetic-logical processing stream and structures processing stream. Among the key areas for future research, development of the mathematical approach to provide a subsequently formalized and automated process of parallelizing sequential algorithms between the CPU and structures processor is highlighted. Among the mathematical models that can be used in future studies there are graph models of algorithms (e.g., dependency graph of a program. Due to the high

  9. Handling multiple metadata streams regarding digital learning material

    NARCIS (Netherlands)

    Roes, J.B.M.; Vuuren, J. van; Verbeij, N.; Nijstad, H.

    2010-01-01

    This paper presents the outcome of a study performed in the Netherlands on handling multiple metadata streams regarding digital learning material. The paper describes the present metadata architecture in the Netherlands, the present suppliers and users of metadata and digital learning materials. It

  10. Data streams: algorithms and applications

    National Research Council Canada - National Science Library

    Muthukrishnan, S

    2005-01-01

    ... massive data sets in general. Researchers in Theoretical Computer Science, Databases, IP Networking and Computer Systems are working on the data stream challenges. This article is an overview and survey of data stream algorithmics and is an updated version of [175]. S. Muthukrishnan Rutgers University, New Brunswick, NJ, USA, muthu@cs...

  11. Comparing the performance of SIMD computers by running large air pollution models

    DEFF Research Database (Denmark)

    Brown, J.; Hansen, Per Christian; Wasniewski, J.

    1996-01-01

    To compare the performance and use of three massively parallel SIMD computers, we implemented a large air pollution model on these computers. Using a realistic large-scale model, we gained detailed insight about the performance of the computers involved when used to solve large-scale scientific...... problems that involve several types of numerical computations. The computers used in our study are the Connection Machines CM-200 and CM-5, and the MasPar MP-2216...

  12. Faster Smith-Waterman database searches with inter-sequence SIMD parallelisation.

    Science.gov (United States)

    Rognes, Torbjørn

    2011-06-01

    The Smith-Waterman algorithm for local sequence alignment is more sensitive than heuristic methods for database searching, but also more time-consuming. The fastest approach to parallelisation with SIMD technology has previously been described by Farrar in 2007. The aim of this study was to explore whether further speed could be gained by other approaches to parallelisation. A faster approach and implementation is described and benchmarked. In the new tool SWIPE, residues from sixteen different database sequences are compared in parallel to one query residue. Using a 375 residue query sequence a speed of 106 billion cell updates per second (GCUPS) was achieved on a dual Intel Xeon X5650 six-core processor system, which is over six times more rapid than software based on Farrar's 'striped' approach. SWIPE was about 2.5 times faster when the programs used only a single thread. For shorter queries, the increase in speed was larger. SWIPE was about twice as fast as BLAST when using the BLOSUM50 score matrix, while BLAST was about twice as fast as SWIPE for the BLOSUM62 matrix. The software is designed for 64 bit Linux on processors with SSSE3. Source code is available from http://dna.uio.no/swipe/ under the GNU Affero General Public License. Efficient parallelisation using SIMD on standard hardware makes it possible to run Smith-Waterman database searches more than six times faster than before. The approach described here could significantly widen the potential application of Smith-Waterman searches. Other applications that require optimal local alignment scores could also benefit from improved performance.

  13. Guidelines for the collection of continuous stream water-temperature data in Alaska

    Science.gov (United States)

    Toohey, Ryan C.; Neal, Edward G.; Solin, Gary L.

    2014-01-01

    Objectives of stream monitoring programs differ considerably among many of the academic, Federal, state, tribal, and non-profit organizations in the state of Alaska. Broad inclusion of stream-temperature monitoring can provide an opportunity for collaboration in the development of a statewide stream-temperature database. Statewide and regional coordination could reduce overall monitoring cost, while providing better analyses at multiple spatial and temporal scales to improve resource decision-making. Increased adoption of standardized protocols and data-quality standards may allow for validation of historical modeling efforts with better projection calibration. For records of stream water temperature to be generally consistent, unbiased, and reproducible, data must be collected and analyzed according to documented protocols. Collection of water-temperature data requires definition of data-quality objectives, good site selection, proper selection of instrumentation, proper installation of sensors, periodic site visits to maintain sensors and download data, pre- and post-deployment verification against an NIST-certified thermometer, potential data corrections, and proper documentation, review, and approval. A study created to develop a quality-assurance project plan, data-quality objectives, and a database management plan that includes procedures for data archiving and dissemination could provide a means to standardize a statewide stream-temperature database in Alaska. Protocols can be modified depending on desired accuracy or specific needs of data collected. This document is intended to guide users in collecting time series water-temperature data in Alaskan streams and draws extensively on the broader protocols already published by the U.S. Geological Survey.

  14. Attending to Multiple Visual Streams: Interactions between Location-Based and Category-Based Attentional Selection

    Science.gov (United States)

    Fagioli, Sabrina; Macaluso, Emiliano

    2009-01-01

    Behavioral studies indicate that subjects are able to divide attention between multiple streams of information at different locations. However, it is still unclear to what extent the observed costs reflect processes specifically associated with spatial attention, versus more general interference due the concurrent monitoring of multiple streams of…

  15. Toward 3D-IPTV: design and implementation of a stereoscopic and multiple-perspective video streaming system

    Science.gov (United States)

    Petrovic, Goran; Farin, Dirk; de With, Peter H. N.

    2008-02-01

    3D-Video systems allow a user to perceive depth in the viewed scene and to display the scene from arbitrary viewpoints interactively and on-demand. This paper presents a prototype implementation of a 3D-video streaming system using an IP network. The architecture of our streaming system is layered, where each information layer conveys a single coded video signal or coded scene-description data. We demonstrate the benefits of a layered architecture with two examples: (a) stereoscopic video streaming, (b) monoscopic video streaming with remote multiple-perspective rendering. Our implementation experiments confirm that prototyping 3D-video streaming systems is possible with today's software and hardware. Furthermore, our current operational prototype demonstrates that highly heterogeneous clients can coexist in the system, ranging from auto-stereoscopic 3D displays to resource-constrained mobile devices.

  16. Faster Smith-Waterman database searches with inter-sequence SIMD parallelisation

    Directory of Open Access Journals (Sweden)

    Rognes Torbjørn

    2011-06-01

    Full Text Available Abstract Background The Smith-Waterman algorithm for local sequence alignment is more sensitive than heuristic methods for database searching, but also more time-consuming. The fastest approach to parallelisation with SIMD technology has previously been described by Farrar in 2007. The aim of this study was to explore whether further speed could be gained by other approaches to parallelisation. Results A faster approach and implementation is described and benchmarked. In the new tool SWIPE, residues from sixteen different database sequences are compared in parallel to one query residue. Using a 375 residue query sequence a speed of 106 billion cell updates per second (GCUPS was achieved on a dual Intel Xeon X5650 six-core processor system, which is over six times more rapid than software based on Farrar's 'striped' approach. SWIPE was about 2.5 times faster when the programs used only a single thread. For shorter queries, the increase in speed was larger. SWIPE was about twice as fast as BLAST when using the BLOSUM50 score matrix, while BLAST was about twice as fast as SWIPE for the BLOSUM62 matrix. The software is designed for 64 bit Linux on processors with SSSE3. Source code is available from http://dna.uio.no/swipe/ under the GNU Affero General Public License. Conclusions Efficient parallelisation using SIMD on standard hardware makes it possible to run Smith-Waterman database searches more than six times faster than before. The approach described here could significantly widen the potential application of Smith-Waterman searches. Other applications that require optimal local alignment scores could also benefit from improved performance.

  17. High-Performance Parallel and Stream Processing of X-ray Microdiffraction Data on Multicores

    International Nuclear Information System (INIS)

    Bauer, Michael A; McIntyre, Stewart; Xie Yuzhen; Biem, Alain; Tamura, Nobumichi

    2012-01-01

    We present the design and implementation of a high-performance system for processing synchrotron X-ray microdiffraction (XRD) data in IBM InfoSphere Streams on multicore processors. We report on the parallel and stream processing techniques that we use to harvest the power of clusters of multicores to analyze hundreds of gigabytes of synchrotron XRD data in order to reveal the microtexture of polycrystalline materials. The timing to process one XRD image using one pipeline is about ten times faster than the best C program at present. With the support of InfoSphere Streams platform, our software is able to be scaled up to operate on clusters of multi-cores for processing multiple images concurrently. This system provides a high-performance processing kernel to achieve near real-time data analysis of image data from synchrotron experiments.

  18. Optimizing Cost of Continuous Overlapping Queries over Data Streams by Filter Adaption

    KAUST Repository

    Xie, Qing

    2016-01-12

    The problem we aim to address is the optimization of cost management for executing multiple continuous queries on data streams, where each query is defined by several filters, each of which monitors certain status of the data stream. Specially the filter can be shared by different queries and expensive to evaluate. The conventional objective for such a problem is to minimize the overall execution cost to solve all queries, by planning the order of filter evaluation in shared strategy. However, in streaming scenario, the characteristics of data items may change in process, which can bring some uncertainty to the outcome of individual filter evaluation, and affect the plan of query execution as well as the overall execution cost. In our work, considering the influence of the uncertain variation of data characteristics, we propose a framework to deal with the dynamic adjustment of filter ordering for query execution on data stream, and focus on the issues of cost management. By incrementally monitoring and analyzing the results of filter evaluation, our proposed approach can be effectively adaptive to the varied stream behavior and adjust the optimal ordering of filter evaluation, so as to optimize the execution cost. In order to achieve satisfactory performance and efficiency, we also discuss the trade-off between the adaptivity of our framework and the overhead incurred by filter adaption. The experimental results on synthetic and two real data sets (traffic and multimedia) show that our framework can effectively reduce and balance the overall query execution cost and keep high adaptivity in streaming scenario.

  19. A real time sorting algorithm to time sort any deterministic time disordered data stream

    Science.gov (United States)

    Saini, J.; Mandal, S.; Chakrabarti, A.; Chattopadhyay, S.

    2017-12-01

    In new generation high intensity high energy physics experiments, millions of free streaming high rate data sources are to be readout. Free streaming data with associated time-stamp can only be controlled by thresholds as there is no trigger information available for the readout. Therefore, these readouts are prone to collect large amount of noise and unwanted data. For this reason, these experiments can have output data rate of several orders of magnitude higher than the useful signal data rate. It is therefore necessary to perform online processing of the data to extract useful information from the full data set. Without trigger information, pre-processing on the free streaming data can only be done with time based correlation among the data set. Multiple data sources have different path delays and bandwidth utilizations and therefore the unsorted merged data requires significant computational efforts for real time manifestation of sorting before analysis. Present work reports a new high speed scalable data stream sorting algorithm with its architectural design, verified through Field programmable Gate Array (FPGA) based hardware simulation. Realistic time based simulated data likely to be collected in an high energy physics experiment have been used to study the performance of the algorithm. The proposed algorithm uses parallel read-write blocks with added memory management and zero suppression features to make it efficient for high rate data-streams. This algorithm is best suited for online data streams with deterministic time disorder/unsorting on FPGA like hardware.

  20. Effective Vectorization with OpenMP 4.5

    Energy Technology Data Exchange (ETDEWEB)

    Huber, Joseph N. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Hernandez, Oscar R. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Lopez, Matthew Graham [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2017-03-01

    This paper describes how the Single Instruction Multiple Data (SIMD) model and its extensions in OpenMP work, and how these are implemented in different compilers. Modern processors are highly parallel computational machines which often include multiple processors capable of executing several instructions in parallel. Understanding SIMD and executing instructions in parallel allows the processor to achieve higher performance without increasing the power required to run it. SIMD instructions can significantly reduce the runtime of code by executing a single operation on large groups of data. The SIMD model is so integral to the processor s potential performance that, if SIMD is not utilized, less than half of the processor is ever actually used. Unfortunately, using SIMD instructions is a challenge in higher level languages because most programming languages do not have a way to describe them. Most compilers are capable of vectorizing code by using the SIMD instructions, but there are many code features important for SIMD vectorization that the compiler cannot determine at compile time. OpenMP attempts to solve this by extending the C++/C and Fortran programming languages with compiler directives that express SIMD parallelism. OpenMP is used to pass hints to the compiler about the code to be executed in SIMD. This is a key resource for making optimized code, but it does not change whether or not the code can use SIMD operations. However, in many cases critical functions are limited by a poor understanding of how SIMD instructions are actually implemented, as SIMD can be implemented through vector instructions or simultaneous multi-threading (SMT). We have found that it is often the case that code cannot be vectorized, or is vectorized poorly, because the programmer does not have sufficient knowledge of how SIMD instructions work.

  1. StreamQRE: Modular Specification and Efficient Evaluation of Quantitative Queries over Streaming Data.

    Science.gov (United States)

    Mamouras, Konstantinos; Raghothaman, Mukund; Alur, Rajeev; Ives, Zachary G; Khanna, Sanjeev

    2017-06-01

    Real-time decision making in emerging IoT applications typically relies on computing quantitative summaries of large data streams in an efficient and incremental manner. To simplify the task of programming the desired logic, we propose StreamQRE, which provides natural and high-level constructs for processing streaming data. Our language has a novel integration of linguistic constructs from two distinct programming paradigms: streaming extensions of relational query languages and quantitative extensions of regular expressions. The former allows the programmer to employ relational constructs to partition the input data by keys and to integrate data streams from different sources, while the latter can be used to exploit the logical hierarchy in the input stream for modular specifications. We first present the core language with a small set of combinators, formal semantics, and a decidable type system. We then show how to express a number of common patterns with illustrative examples. Our compilation algorithm translates the high-level query into a streaming algorithm with precise complexity bounds on per-item processing time and total memory footprint. We also show how to integrate approximation algorithms into our framework. We report on an implementation in Java, and evaluate it with respect to existing high-performance engines for processing streaming data. Our experimental evaluation shows that (1) StreamQRE allows more natural and succinct specification of queries compared to existing frameworks, (2) the throughput of our implementation is higher than comparable systems (for example, two-to-four times greater than RxJava), and (3) the approximation algorithms supported by our implementation can lead to substantial memory savings.

  2. Accelerating finite-rate chemical kinetics with coprocessors: Comparing vectorization methods on GPUs, MICs, and CPUs

    Science.gov (United States)

    Stone, Christopher P.; Alferman, Andrew T.; Niemeyer, Kyle E.

    2018-05-01

    Accurate and efficient methods for solving stiff ordinary differential equations (ODEs) are a critical component of turbulent combustion simulations with finite-rate chemistry. The ODEs governing the chemical kinetics at each mesh point are decoupled by operator-splitting allowing each to be solved concurrently. An efficient ODE solver must then take into account the available thread and instruction-level parallelism of the underlying hardware, especially on many-core coprocessors, as well as the numerical efficiency. A stiff Rosenbrock and a nonstiff Runge-Kutta ODE solver are both implemented using the single instruction, multiple thread (SIMT) and single instruction, multiple data (SIMD) paradigms within OpenCL. Both methods solve multiple ODEs concurrently within the same instruction stream. The performance of these parallel implementations was measured on three chemical kinetic models of increasing size across several multicore and many-core platforms. Two separate benchmarks were conducted to clearly determine any performance advantage offered by either method. The first benchmark measured the run-time of evaluating the right-hand-side source terms in parallel and the second benchmark integrated a series of constant-pressure, homogeneous reactors using the Rosenbrock and Runge-Kutta solvers. The right-hand-side evaluations with SIMD parallelism on the host multicore Xeon CPU and many-core Xeon Phi co-processor performed approximately three times faster than the baseline multithreaded C++ code. The SIMT parallel model on the host and Phi was 13%-35% slower than the baseline while the SIMT model on the NVIDIA Kepler GPU provided approximately the same performance as the SIMD model on the Phi. The runtimes for both ODE solvers decreased significantly with the SIMD implementations on the host CPU (2.5-2.7 ×) and Xeon Phi coprocessor (4.7-4.9 ×) compared to the baseline parallel code. The SIMT implementations on the GPU ran 1.5-1.6 times faster than the baseline

  3. Responses of stream microbes to multiple anthropogenic stressors in a mesocosm study.

    Science.gov (United States)

    Nuy, Julia K; Lange, Anja; Beermann, Arne J; Jensen, Manfred; Elbrecht, Vasco; Röhl, Oliver; Peršoh, Derek; Begerow, Dominik; Leese, Florian; Boenigk, Jens

    2018-08-15

    Stream ecosystems are affected by multiple anthropogenic stressors worldwide. Even though effects of many single stressors are comparatively well studied, the effects of multiple stressors are difficult to predict. In particular bacteria and protists, which are responsible for the majority of ecosystem respiration and element flows, are infrequently studied with respect to multiple stressors responses. We conducted a stream mesocosm experiment to characterize the responses of single and multiple stressors on microbiota. Two functionally important stream habitats, leaf litter and benthic phototrophic rock biofilms, were exposed to three stressors in a full factorial design: fine sediment deposition, increased chloride concentration (salinization) and reduced flow velocity. We analyzed the microbial composition in the two habitat types of the mesocosms using an amplicon sequencing approach. Community analysis on different taxonomic levels as well as principle component analyses (PCoAs) based on realtive abundances of operational taxonomic units (OTUs) showed treatment specific shifts in the eukaryotic biofilm community. Analysis of variance (ANOVA) revealed that Bacillariophyta responded positively salinity and sediment increase, while the relative read abundance of chlorophyte taxa decreased. The combined effects of multiple stressors were mainly antagonistic. Therefore, the community composition in multiply stressed environments resembled the composition of the unstressed control community in terms of OTU occurrence and relative abundances. Copyright © 2018 The Authors. Published by Elsevier B.V. All rights reserved.

  4. Efficient, symmetry-driven SIMD access patterns for 3D PET image reconstruction applicable for CPUs and GPUs

    Energy Technology Data Exchange (ETDEWEB)

    Scheins, J.J.; Garcia Lucio, L.F.; Herzog, H.; Shah, N.J. [Forschungszentrum Juelich GmbH (Germany). Inst. of Neuroscience and Medicine (INM-4)

    2011-07-01

    Fully 3D PET image reconstruction still remains a challenging computational task due to the tremendous number of registered Lines-of-Response. Typically, billions of geometrical weights have to be repeatedly calculated and evaluated for iterative algorithms. In this context, the reconstruction software PRESTO (PET REconstruction Software TOolkit) provides accurate geometrical weighting schemes for the forward projection and backward projection, e.g. Volume-of-Intersection, while using all measured LORs separately. PRESTO exploits redundancies to realise a strongly compressed, memory-resident system matrix. Consequently, the needed time to calculate matrix weights no longer influences the reconstruction time. Very high compression factors (>300) are achieved by using unconventional non-cartesian voxel patterns. However, in the original implementation the addressing of matrix weights, projection values and voxel values happens in disfavoured memory access patterns. This causes severe computational inefficiencies due to the limited memory bandwidth using CPUs. In this work, the image data and projection data in memory as well as the order of mathematical operations have been completely re-organised to provide an optimal merit for the Single Instruction Multiple Data (SIMD) approach. This reorganisation is directly driven by the induced symmetries of PRESTO. A global speedup factor of 15 for has been achieved for the CPU-based implementation while obtaining identical results. In addition, a GPU-based implementation using CUDA on Nvidia TESLA C1060/S1070 hardware provides another speed up factor of 4 compared to single core CPU processing. (orig.)

  5. Efficient, symmetry-driven SIMD access patterns for 3D PET image reconstruction applicable for CPUs and GPUs

    International Nuclear Information System (INIS)

    Scheins, J.J.; Garcia Lucio, L.F.; Herzog, H.; Shah, N.J.

    2011-01-01

    Fully 3D PET image reconstruction still remains a challenging computational task due to the tremendous number of registered Lines-of-Response. Typically, billions of geometrical weights have to be repeatedly calculated and evaluated for iterative algorithms. In this context, the reconstruction software PRESTO (PET REconstruction Software TOolkit) provides accurate geometrical weighting schemes for the forward projection and backward projection, e.g. Volume-of-Intersection, while using all measured LORs separately. PRESTO exploits redundancies to realise a strongly compressed, memory-resident system matrix. Consequently, the needed time to calculate matrix weights no longer influences the reconstruction time. Very high compression factors (>300) are achieved by using unconventional non-cartesian voxel patterns. However, in the original implementation the addressing of matrix weights, projection values and voxel values happens in disfavoured memory access patterns. This causes severe computational inefficiencies due to the limited memory bandwidth using CPUs. In this work, the image data and projection data in memory as well as the order of mathematical operations have been completely re-organised to provide an optimal merit for the Single Instruction Multiple Data (SIMD) approach. This reorganisation is directly driven by the induced symmetries of PRESTO. A global speedup factor of 15 for has been achieved for the CPU-based implementation while obtaining identical results. In addition, a GPU-based implementation using CUDA on Nvidia TESLA C1060/S1070 hardware provides another speed up factor of 4 compared to single core CPU processing. (orig.)

  6. Data Stream Classification Based on the Gamma Classifier

    Directory of Open Access Journals (Sweden)

    Abril Valeria Uriarte-Arcia

    2015-01-01

    Full Text Available The ever increasing data generation confronts us with the problem of handling online massive amounts of information. One of the biggest challenges is how to extract valuable information from these massive continuous data streams during single scanning. In a data stream context, data arrive continuously at high speed; therefore the algorithms developed to address this context must be efficient regarding memory and time management and capable of detecting changes over time in the underlying distribution that generated the data. This work describes a novel method for the task of pattern classification over a continuous data stream based on an associative model. The proposed method is based on the Gamma classifier, which is inspired by the Alpha-Beta associative memories, which are both supervised pattern recognition models. The proposed method is capable of handling the space and time constrain inherent to data stream scenarios. The Data Streaming Gamma classifier (DS-Gamma classifier implements a sliding window approach to provide concept drift detection and a forgetting mechanism. In order to test the classifier, several experiments were performed using different data stream scenarios with real and synthetic data streams. The experimental results show that the method exhibits competitive performance when compared to other state-of-the-art algorithms.

  7. Analytic Strategies of Streaming Data for eHealth.

    Science.gov (United States)

    Yoon, Sunmoo

    2016-01-01

    New analytic strategies for streaming big data from wearable devices and social media are emerging in ehealth. We face challenges to find meaningful patterns from big data because researchers face difficulties to process big volume of streaming data using traditional processing applications.1 This introductory 180 minutes tutorial offers hand-on instruction on analytics2 (e.g., topic modeling, social network analysis) of streaming data. This tutorial aims to provide practical strategies of information on reducing dimensionality using examples of big data. This tutorial will highlight strategies of incorporating domain experts and a comprehensive approach to streaming social media data.

  8. Human Factors in Streaming Data Analysis: Challenges and Opportunities for Information Visualization: Human Factors in Streaming Data Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Dasgupta, Aritra [Pacific Northwest National Laboratory, Richland Washington USA; Arendt, Dustin L. [Pacific Northwest National Laboratory, Richland Washington USA; Franklin, Lyndsey R. [Pacific Northwest National Laboratory, Richland Washington USA; Wong, Pak Chung [Pacific Northwest National Laboratory, Richland Washington USA; Cook, Kristin A. [Pacific Northwest National Laboratory, Richland Washington USA

    2017-09-01

    Real-world systems change continuously and across domains like traffic monitoring, cyber security, etc., such changes occur within short time scales. This leads to a streaming data problem and produces unique challenges for the human in the loop, as analysts have to ingest and make sense of dynamic patterns in real time. In this paper, our goal is to study how the state-of-the-art in streaming data visualization handles these challenges and reflect on the gaps and opportunities. To this end, we have three contributions: i) problem characterization for identifying domain-specific goals and challenges for handling streaming data, ii) a survey and analysis of the state-of-the-art in streaming data visualization research with a focus on the visualization design space, and iii) reflections on the perceptually motivated design challenges and potential research directions for addressing them.

  9. Data Stream Clustering With Affinity Propagation

    KAUST Repository

    Zhang, Xiangliang

    2014-07-09

    Data stream clustering provides insights into the underlying patterns of data flows. This paper focuses on selecting the best representatives from clusters of streaming data. There are two main challenges: how to cluster with the best representatives and how to handle the evolving patterns that are important characteristics of streaming data with dynamic distributions. We employ the Affinity Propagation (AP) algorithm presented in 2007 by Frey and Dueck for the first challenge, as it offers good guarantees of clustering optimality for selecting exemplars. The second challenging problem is solved by change detection. The presented StrAP algorithm combines AP with a statistical change point detection test; the clustering model is rebuilt whenever the test detects a change in the underlying data distribution. Besides the validation on two benchmark data sets, the presented algorithm is validated on a real-world application, monitoring the data flow of jobs submitted to the EGEE grid.

  10. Data Stream Clustering With Affinity Propagation

    KAUST Repository

    Zhang, Xiangliang; Furtlehner, Cyril; Germain-Renaud, Cecile; Sebag, Michele

    2014-01-01

    Data stream clustering provides insights into the underlying patterns of data flows. This paper focuses on selecting the best representatives from clusters of streaming data. There are two main challenges: how to cluster with the best representatives and how to handle the evolving patterns that are important characteristics of streaming data with dynamic distributions. We employ the Affinity Propagation (AP) algorithm presented in 2007 by Frey and Dueck for the first challenge, as it offers good guarantees of clustering optimality for selecting exemplars. The second challenging problem is solved by change detection. The presented StrAP algorithm combines AP with a statistical change point detection test; the clustering model is rebuilt whenever the test detects a change in the underlying data distribution. Besides the validation on two benchmark data sets, the presented algorithm is validated on a real-world application, monitoring the data flow of jobs submitted to the EGEE grid.

  11. An Approach for Removing Redundant Data from RFID Data Streams

    Science.gov (United States)

    Mahdin, Hairulnizam; Abawajy, Jemal

    2011-01-01

    Radio frequency identification (RFID) systems are emerging as the primary object identification mechanism, especially in supply chain management. However, RFID naturally generates a large amount of duplicate readings. Removing these duplicates from the RFID data stream is paramount as it does not contribute new information to the system and wastes system resources. Existing approaches to deal with this problem cannot fulfill the real time demands to process the massive RFID data stream. We propose a data filtering approach that efficiently detects and removes duplicate readings from RFID data streams. Experimental results show that the proposed approach offers a significant improvement as compared to the existing approaches. PMID:22163730

  12. An Approach for Removing Redundant Data from RFID Data Streams

    Directory of Open Access Journals (Sweden)

    Hairulnizam Mahdin

    2011-10-01

    Full Text Available Radio frequency identification (RFID systems are emerging as the primary object identification mechanism, especially in supply chain management. However, RFID naturally generates a large amount of duplicate readings. Removing these duplicates from the RFID data stream is paramount as it does not contribute new information to the system and wastes system resources. Existing approaches to deal with this problem cannot fulfill the real time demands to process the massive RFID data stream. We propose a data filtering approach that efficiently detects and removes duplicate readings from RFID data streams. Experimental results show that the proposed approach offers a significant improvement as compared to the existing approaches.

  13. A Multiple Streams analysis of the decisions to fund gender-neutral HPV vaccination in Canada.

    Science.gov (United States)

    Shapiro, Gilla K; Guichon, Juliet; Prue, Gillian; Perez, Samara; Rosberger, Zeev

    2017-07-01

    In Canada, the human papillomavirus (HPV) vaccine is licensed and recommended for females and males. Although all Canadian jurisdictions fund school-based HPV vaccine programs for girls, only six jurisdictions fund school-based HPV vaccination for boys. The research aimed to analyze the factors that underpin government decisions to fund HPV vaccine for boys using a theoretical policy model, Kingdon's Multiple Streams framework. This approach assesses policy development by examining three concurrent, but independent, streams that guide analysis: Problem Stream, Policy Stream, and Politics Stream. Analysis from the Problem Stream highlights that males are affected by HPV-related diseases and are involved in transmitting HPV infection to their sexual partners. Policy Stream analysis makes clear that while the inclusion of males in HPV vaccine programs is suitable, equitable, and acceptable; there is debate regarding cost-effectiveness. Politics Stream analysis identifies the perspectives of six different stakeholder groups and highlights the contribution of government officials at the provincial and territorial level. Kingdon's Multiple Streams framework helps clarify the opportunities and barriers for HPV vaccine policy change. This analysis identified that the interpretation of cost-effectiveness models and advocacy of stakeholders such as citizen-advocates and HPV-affected politicians have been particularly important in galvanizing policy change. Copyright © 2017 Elsevier Inc. All rights reserved.

  14. A survey on Big Data Stream Mining

    African Journals Online (AJOL)

    pc

    2018-03-05

    Mar 5, 2018 ... huge amount of stream like telecommunication systems. So, there ... streams have many challenges for data mining algorithm design like using of ..... A. Bifet and R. Gavalda, "Learning from Time-Changing Data with. Adaptive ...

  15. Variation partitioning of diatom species data matrices: Understanding the influence of multiple factors on benthic diatom communities in tropical streams

    Energy Technology Data Exchange (ETDEWEB)

    Bere, Taurai, E-mail: tbere2015@gmail.com; Mangadze, Tinotenda; Mwedzi, Tongai

    2016-10-01

    Elucidating the confounding influence of multiple environmental factors on benthic diatom communities is important in developing water quality predictive models for better guidance of stream management efforts. The objective of this study was to explore the relative impact of metal pollution and hydromorphological alterations in, addition to nutrient enrichment and organic pollution, on diatom taxonomic composition with the view to improve stream diatom-based water quality inference models. Samples were collected twice at 20 sampling stations in the tropical Manyame Catchment, Zimbabwe. Diatom, macroinvertebrate communities and environmental factors were sampled and analysed. The variations in diatom community composition explained by different categories of environmental factors were analysed using canonical correspondence analysis using variance partitioning (partial CCA). The following variations were explained by the different predictor matrices: nutrient levels and organic pollution - 10.4%, metal pollution - 8.3% and hydromorphological factors - 7.9%. Thus, factors other than nutrient levels and organic pollution explain additional significant variation in these diatom communities. Development of diatom-based stream water quality inference models that incorporate metal pollution and hydromorphological alterations, where these are key issues, is thus deemed necessary. - Highlights: • Confounding influences of multiple environmental factors on diatom communities are elucidated. • Variation explained: nutrients + organic pollution - 10.4%, metals - 8.3% and hydromorphological factors - 7.9%. • Calibration of existing or development of new indices may be necessary.

  16. Variation partitioning of diatom species data matrices: Understanding the influence of multiple factors on benthic diatom communities in tropical streams

    International Nuclear Information System (INIS)

    Bere, Taurai; Mangadze, Tinotenda; Mwedzi, Tongai

    2016-01-01

    Elucidating the confounding influence of multiple environmental factors on benthic diatom communities is important in developing water quality predictive models for better guidance of stream management efforts. The objective of this study was to explore the relative impact of metal pollution and hydromorphological alterations in, addition to nutrient enrichment and organic pollution, on diatom taxonomic composition with the view to improve stream diatom-based water quality inference models. Samples were collected twice at 20 sampling stations in the tropical Manyame Catchment, Zimbabwe. Diatom, macroinvertebrate communities and environmental factors were sampled and analysed. The variations in diatom community composition explained by different categories of environmental factors were analysed using canonical correspondence analysis using variance partitioning (partial CCA). The following variations were explained by the different predictor matrices: nutrient levels and organic pollution - 10.4%, metal pollution - 8.3% and hydromorphological factors - 7.9%. Thus, factors other than nutrient levels and organic pollution explain additional significant variation in these diatom communities. Development of diatom-based stream water quality inference models that incorporate metal pollution and hydromorphological alterations, where these are key issues, is thus deemed necessary. - Highlights: • Confounding influences of multiple environmental factors on diatom communities are elucidated. • Variation explained: nutrients + organic pollution - 10.4%, metals - 8.3% and hydromorphological factors - 7.9%. • Calibration of existing or development of new indices may be necessary.

  17. What Can Hierarchies Do for Data Streams?

    DEFF Research Database (Denmark)

    Yin, Xuepeng; Pedersen, Torben Bach

    2007-01-01

    Abstract. Much effort has been put into building data streams management systems for querying data streams. However, the query languages have mostly been SQL-based and aimed for low-level analysis of base data; therefore, there has been little work on supporting OLAP-like queries that provide rea...

  18. Towards adaptive, streaming analysis of x-ray tomography data

    Energy Technology Data Exchange (ETDEWEB)

    Thomas, Mathew; Kleese van Dam, Kerstin; Marshall, Matthew J.; Kuprat, Andrew P.; Carson, James P.; Lansing, Carina S.; Guillen, Zoe C.; Miller, Erin A.; Lanekoff, Ingela; Laskin, Julia

    2015-03-04

    Temporal and spatial resolution of chemical imaging methodologies such as x-ray tomography are rapidly increasing, leading to more complex experimental procedures and fast growing data volumes. Automated analysis pipelines and big data analytics are becoming essential to effectively evaluate the results of such experiments. Offering those data techniques in an adaptive, streaming environment can further substantially improve the scientific discovery process, by enabling experimental control and steering based on the evaluation of emerging phenomena as they are observed by the experiment. Pacific Northwest National Laboratory (PNNL)’ Chemical Imaging Initiative (CII - http://imaging.pnnl.gov/ ) has worked since 2011 towards developing a framework that allows users to rapidly compose and customize high throughput experimental analysis pipelines for multiple instrument types. The framework, named ‘Rapid Experimental Analysis’ (REXAN) Framework [1], is based on the idea of reusable component libraries and utilizes the PNNL developed collaborative data management and analysis environment ‘Velo’, to provide a user friendly analysis and data management environment for experimental facilities. This article will, discuss the capabilities established for X-Ray tomography, discuss lessons learned, and provide an overview of our more recent work in the Analysis in Motion Initiative (AIM - http://aim.pnnl.gov/ ) at PNNL to provide REXAN capabilities in a streaming environment.

  19. Prediction and explanation over DL-Lite data streams

    CSIR Research Space (South Africa)

    Klarman, S

    2013-12-01

    Full Text Available the popular DL-Lite family, and study the logic foundations of prediction and explanation over DL-Lite data streams, i.e., reasoning from finite segments of streaming data to conjectures about the content of the streams in the future or in the past. We propose...

  20. Alignment data streams for the ATLAS inner detector

    CERN Document Server

    Pinto, B; Pereira, P; Elsing, M; Hawkings, R; Schieck, J; García, S; Schaffer, A; Ma, H; Anjos, A

    2008-01-01

    The ATLAS experiment uses a complex trigger strategy to be able to reduce the Event Filter rate output, down to a level that allows the storage and processing of these data. These concepts are described in the ATLAS Computing Model which embraces Grid paradigm. The output coming from the Event Filter consists of four main streams: physical stream, express stream, calibration stream, and diagnostic stream. The calibration stream will be transferred to the Tier-0 facilities that will provide the prompt reconstruction of this stream with a minimum latency of 8 hours, producing calibration constants of sufficient quality to allow a first-pass processing. The Inner Detector community is developing and testing an independent common calibration stream selected at the Event Filter after track reconstruction. It is composed of raw data, in byte-stream format, contained in Readout Buffers (ROBs) with hit information of the selected tracks, and it will be used to derive and update a set of calibration and alignment cons...

  1. Streaming Model Based Volume Ray Casting Implementation for Cell Broadband Engine

    Directory of Open Access Journals (Sweden)

    Jusub Kim

    2009-01-01

    Full Text Available Interactive high quality volume rendering is becoming increasingly more important as the amount of more complex volumetric data steadily grows. While a number of volumetric rendering techniques have been widely used, ray casting has been recognized as an effective approach for generating high quality visualization. However, for most users, the use of ray casting has been limited to datasets that are very small because of its high demands on computational power and memory bandwidth. However the recent introduction of the Cell Broadband Engine (Cell B.E. processor, which consists of 9 heterogeneous cores designed to handle extremely demanding computations with large streams of data, provides an opportunity to put the ray casting into practical use. In this paper, we introduce an efficient parallel implementation of volume ray casting on the Cell B.E. The implementation is designed to take full advantage of the computational power and memory bandwidth of the Cell B.E. using an intricate orchestration of the ray casting computation on the available heterogeneous resources. Specifically, we introduce streaming model based schemes and techniques to efficiently implement acceleration techniques for ray casting on Cell B.E. In addition to ensuring effective SIMD utilization, our method provides two key benefits: there is no cost for empty space skipping and there is no memory bottleneck on moving volumetric data for processing. Our experimental results show that we can interactively render practical datasets on a single Cell B.E. processor.

  2. Continuous sampling from distributed streams

    DEFF Research Database (Denmark)

    Graham, Cormode; Muthukrishnan, S.; Yi, Ke

    2012-01-01

    A fundamental problem in data management is to draw and maintain a sample of a large data set, for approximate query answering, selectivity estimation, and query planning. With large, streaming data sets, this problem becomes particularly difficult when the data is shared across multiple distribu......A fundamental problem in data management is to draw and maintain a sample of a large data set, for approximate query answering, selectivity estimation, and query planning. With large, streaming data sets, this problem becomes particularly difficult when the data is shared across multiple...... distributed sites. The main challenge is to ensure that a sample is drawn uniformly across the union of the data while minimizing the communication needed to run the protocol on the evolving data. At the same time, it is also necessary to make the protocol lightweight, by keeping the space and time costs low...... for each participant. In this article, we present communication-efficient protocols for continuously maintaining a sample (both with and without replacement) from k distributed streams. These apply to the case when we want a sample from the full streams, and to the sliding window cases of only the W most...

  3. Interaction sorting method for molecular dynamics on multi-core SIMD CPU architecture.

    Science.gov (United States)

    Matvienko, Sergey; Alemasov, Nikolay; Fomin, Eduard

    2015-02-01

    Molecular dynamics (MD) is widely used in computational biology for studying binding mechanisms of molecules, molecular transport, conformational transitions, protein folding, etc. The method is computationally expensive; thus, the demand for the development of novel, much more efficient algorithms is still high. Therefore, the new algorithm designed in 2007 and called interaction sorting (IS) clearly attracted interest, as it outperformed the most efficient MD algorithms. In this work, a new IS modification is proposed which allows the algorithm to utilize SIMD processor instructions. This paper shows that the improvement provides an additional gain in performance, 9% to 45% in comparison to the original IS method.

  4. Mining Building Metadata by Data Stream Comparison

    DEFF Research Database (Denmark)

    Holmegaard, Emil; Kjærgaard, Mikkel Baun

    2016-01-01

    to handle data streams with only slightly similar patterns. We have evaluated Metafier with points and data from one building located in Denmark. We have evaluated Metafier with 903 points, and the overall accuracy, with only 3 known examples, was 94.71%. Furthermore we found that using DTW for mining...... ways to annotate sensor and actuation points. This makes it difficult to create intuitive queries for retrieving data streams from points. Another problem is the amount of insufficient or missing metadata. We introduce Metafier, a tool for extracting metadata from comparing data streams. Metafier...... enables a semi-automatic labeling of metadata to building instrumentation. Metafier annotates points with metadata by comparing the data from a set of validated points with unvalidated points. Metafier has three different algorithms to compare points with based on their data. The three algorithms...

  5. A Context-Aware Adaptive Streaming Media Distribution System in a Heterogeneous Network with Multiple Terminals

    Directory of Open Access Journals (Sweden)

    Yepeng Ni

    2016-01-01

    Full Text Available We consider the problem of streaming media transmission in a heterogeneous network from a multisource server to home multiple terminals. In wired network, the transmission performance is limited by network state (e.g., the bandwidth variation, jitter, and packet loss. In wireless network, the multiple user terminals can cause bandwidth competition. Thus, the streaming media distribution in a heterogeneous network becomes a severe challenge which is critical for QoS guarantee. In this paper, we propose a context-aware adaptive streaming media distribution system (CAASS, which implements the context-aware module to perceive the environment parameters and use the strategy analysis (SA module to deduce the most suitable service level. This approach is able to improve the video quality for guarantying streaming QoS. We formulate the optimization problem of QoS relationship with the environment parameters based on the QoS testing algorithm for IPTV in ITU-T G.1070. We evaluate the performance of the proposed CAASS through 12 types of experimental environments using a prototype system. Experimental results show that CAASS can dynamically adjust the service level according to the environment variation (e.g., network state and terminal performances and outperforms the existing streaming approaches in adaptive streaming media distribution according to peak signal-to-noise ratio (PSNR.

  6. AMIDST: Analysis of MassIve Data STreams

    DEFF Research Database (Denmark)

    Masegosa, Andres; Martinez, Ana Maria; Borchani, Hanen

    2015-01-01

    The Analysis of MassIve Data STreams (AMIDST) Java toolbox provides a collection of scalable and parallel algorithms for inference and learning of hybrid Bayesian networks from data streams. The toolbox, available at http://amidst.github.io/toolbox/ under the Apache Software License version 2.......0, also efficiently leverages existing functionalities and algorithms by interfacing to software tools such as HUGIN and MOA....

  7. Scenario driven data modelling: a method for integrating diverse sources of data and data streams

    Science.gov (United States)

    2011-01-01

    Background Biology is rapidly becoming a data intensive, data-driven science. It is essential that data is represented and connected in ways that best represent its full conceptual content and allows both automated integration and data driven decision-making. Recent advancements in distributed multi-relational directed graphs, implemented in the form of the Semantic Web make it possible to deal with complicated heterogeneous data in new and interesting ways. Results This paper presents a new approach, scenario driven data modelling (SDDM), that integrates multi-relational directed graphs with data streams. SDDM can be applied to virtually any data integration challenge with widely divergent types of data and data streams. In this work, we explored integrating genetics data with reports from traditional media. SDDM was applied to the New Delhi metallo-beta-lactamase gene (NDM-1), an emerging global health threat. The SDDM process constructed a scenario, created a RDF multi-relational directed graph that linked diverse types of data to the Semantic Web, implemented RDF conversion tools (RDFizers) to bring content into the Sematic Web, identified data streams and analytical routines to analyse those streams, and identified user requirements and graph traversals to meet end-user requirements. Conclusions We provided an example where SDDM was applied to a complex data integration challenge. The process created a model of the emerging NDM-1 health threat, identified and filled gaps in that model, and constructed reliable software that monitored data streams based on the scenario derived multi-relational directed graph. The SDDM process significantly reduced the software requirements phase by letting the scenario and resulting multi-relational directed graph define what is possible and then set the scope of the user requirements. Approaches like SDDM will be critical to the future of data intensive, data-driven science because they automate the process of converting

  8. StreamSqueeze: a dynamic stream visualization for monitoring of event data

    Science.gov (United States)

    Mansmann, Florian; Krstajic, Milos; Fischer, Fabian; Bertini, Enrico

    2012-01-01

    While in clear-cut situations automated analytical solution for data streams are already in place, only few visual approaches have been proposed in the literature for exploratory analysis tasks on dynamic information. However, due to the competitive or security-related advantages that real-time information gives in domains such as finance, business or networking, we are convinced that there is a need for exploratory visualization tools for data streams. Under the conditions that new events have higher relevance and that smooth transitions enable traceability of items, we propose a novel dynamic stream visualization called StreamSqueeze. In this technique the degree of interest of recent items is expressed through an increase in size and thus recent events can be shown with more details. The technique has two main benefits: First, the layout algorithm arranges items in several lists of various sizes and optimizes the positions within each list so that the transition of an item from one list to the other triggers least visual changes. Second, the animation scheme ensures that for 50 percent of the time an item has a static screen position where reading is most effective and then continuously shrinks and moves to the its next static position in the subsequent list. To demonstrate the capability of our technique, we apply it to large and high-frequency news and syslog streams and show how it maintains optimal stability of the layout under the conditions given above.

  9. Analyzing indicators of stream health for Minnesota streams

    Science.gov (United States)

    Singh, U.; Kocian, M.; Wilson, B.; Bolton, A.; Nieber, J.; Vondracek, B.; Perry, J.; Magner, J.

    2005-01-01

    Recent research has emphasized the importance of using physical, chemical, and biological indicators of stream health for diagnosing impaired watersheds and their receiving water bodies. A multidisciplinary team at the University of Minnesota is carrying out research to develop a stream classification system for Total Maximum Daily Load (TMDL) assessment. Funding for this research is provided by the United States Environmental Protection Agency and the Minnesota Pollution Control Agency. One objective of the research study involves investigating the relationships between indicators of stream health and localized stream characteristics. Measured data from Minnesota streams collected by various government and non-government agencies and research institutions have been obtained for the research study. Innovative Geographic Information Systems tools developed by the Environmental Science Research Institute and the University of Texas are being utilized to combine and organize the data. Simple linear relationships between index of biological integrity (IBI) and channel slope, two-year stream flow, and drainage area are presented for the Redwood River and the Snake River Basins. Results suggest that more rigorous techniques are needed to successfully capture trends in IBI scores. Additional analyses will be done using multiple regression, principal component analysis, and clustering techniques. Uncovering key independent variables and understanding how they fit together to influence stream health are critical in the development of a stream classification for TMDL assessment.

  10. Event Streams Clustering Using Machine Learning Techniques

    Directory of Open Access Journals (Sweden)

    Hanen Bouali

    2015-10-01

    Full Text Available Data streams are usually of unbounded lengths which push users to consider only recent observations by focusing on a time window, and ignore past data. However, in many real world applications, past data must be taken in consideration to guarantee the efficiency, the performance of decision making and to handle data streams evolution over time. In order to build a selectively history to track the underlying event streams changes, we opt for the continuously data of the sliding window which increases the time window based on changes over historical data. In this paper, to have the ability to access to historical data without requiring any significant storage or multiple passes over the data. In this paper, we propose a new algorithm for clustering multiple data streams using incremental support vector machine and data representative points’ technique. The algorithm uses a sliding window model for the most recent clustering results and data representative points to model the old data clustering results. Our experimental results on electromyography signal show a better clustering than other present in the literature

  11. A High-Performance Parallel FDTD Method Enhanced by Using SSE Instruction Set

    Directory of Open Access Journals (Sweden)

    Dau-Chyrh Chang

    2012-01-01

    Full Text Available We introduce a hardware acceleration technique for the parallel finite difference time domain (FDTD method using the SSE (streaming (single instruction multiple data SIMD extensions instruction set. The implementation of SSE instruction set to parallel FDTD method has achieved the significant improvement on the simulation performance. The benchmarks of the SSE acceleration on both the multi-CPU workstation and computer cluster have demonstrated the advantages of (vector arithmetic logic unit VALU acceleration over GPU acceleration. Several engineering applications are employed to demonstrate the performance of parallel FDTD method enhanced by SSE instruction set.

  12. Alignment data streams for the ATLAS inner detector

    International Nuclear Information System (INIS)

    Pinto, B; Amorim, A; Pereira, P; Elsing, M; Hawkings, R; Schieck, J; Garcia, S; Schaffer, A; Ma, H; Anjos, A

    2008-01-01

    The ATLAS experiment uses a complex trigger strategy to be able to reduce the Event Filter rate output, down to a level that allows the storage and processing of these data. These concepts are described in the ATLAS Computing Model which embraces Grid paradigm. The output coming from the Event Filter consists of four main streams: physical stream, express stream, calibration stream, and diagnostic stream. The calibration stream will be transferred to the Tier-0 facilities that will provide the prompt reconstruction of this stream with a minimum latency of 8 hours, producing calibration constants of sufficient quality to allow a first-pass processing. The Inner Detector community is developing and testing an independent common calibration stream selected at the Event Filter after track reconstruction. It is composed of raw data, in byte-stream format, contained in Readout Buffers (ROBs) with hit information of the selected tracks, and it will be used to derive and update a set of calibration and alignment constants. This option was selected because it makes use of the Byte Stream Converter infrastructure and possibly gives better bandwidth usage and storage optimization. Processing is done using specialized algorithms running in the Athena framework in dedicated Tier-0 resources, and the alignment constants will be stored and distributed using the COOL conditions database infrastructure. This work is addressing in particular the alignment requirements, the needs for track and hit selection, and the performance issues

  13. Alignment data stream for the ATLAS inner detector

    International Nuclear Information System (INIS)

    Pinto, B

    2010-01-01

    The ATLAS experiment uses a complex trigger strategy to be able to achieve the necessary Event Filter rate output, making possible to optimize the storage and processing needs of these data. These needs are described in the ATLAS Computing Model, which embraces Grid concepts. The output coming from the Event Filter will consist of three main streams: a primary stream, the express stream and the calibration stream. The calibration stream will be transferred to the Tier-0 facilities which will allow the prompt reconstruction of this stream with an admissible latency of 8 hours, producing calibration constants of sufficient quality to permit a first-pass processing. An independent calibration stream is developed and tested, which selects tracks at the level-2 trigger (LVL2) after the reconstruction. The stream is composed of raw data, in byte-stream format, and contains only information of the relevant parts of the detector, in particular the hit information of the selected tracks. This leads to a significantly improved bandwidth usage and storage capability. The stream will be used to derive and update the calibration and alignment constants if necessary every 24h. Processing is done using specialized algorithms running in Athena framework in dedicated Tier-0 resources, and the alignment constants will be stored and distributed using the COOL conditions database infrastructure. The work is addressing in particular the alignment requirements, the needs for track and hit selection, timing and bandwidth issues.

  14. A catchment scale evaluation of multiple stressor effects in headwater streams.

    Science.gov (United States)

    Rasmussen, Jes J; McKnight, Ursula S; Loinaz, Maria C; Thomsen, Nanna I; Olsson, Mikael E; Bjerg, Poul L; Binning, Philip J; Kronvang, Brian

    2013-01-01

    Mitigation activities to improve water quality and quantity in streams as well as stream management and restoration efforts are conducted in the European Union aiming to improve the chemical, physical and ecological status of streams. Headwater streams are often characterised by impairment of hydromorphological, chemical, and ecological conditions due to multiple anthropogenic impacts. However, they are generally disregarded as water bodies for mitigation activities in the European Water Framework Directive despite their importance for supporting a higher ecological quality in higher order streams. We studied 11 headwater streams in the Hove catchment in the Copenhagen region. All sites had substantial physical habitat and water quality impairments due to anthropogenic influence (intensive agriculture, urban settlements, contaminated sites and low base-flow due to water abstraction activities in the catchment). We aimed to identify the dominating anthropogenic stressors at the catchment scale causing ecological impairment of benthic macroinvertebrate communities and provide a rank-order of importance that could help in prioritising mitigation activities. We identified numerous chemical and hydromorphological impacts of which several were probably causing major ecological impairments, but we were unable to provide a robust rank-ordering of importance suggesting that targeted mitigation efforts on single anthropogenic stressors in the catchment are unlikely to have substantial effects on the ecological quality in these streams. The SPEcies At Risk (SPEAR) index explained most of the variability in the macroinvertebrate community structure, and notably, SPEAR index scores were often very low (<10% SPEAR abundance). An extensive re-sampling of a subset of the streams provided evidence that especially insecticides were probably essential contributors to the overall ecological impairment of these streams. Our results suggest that headwater streams should be considered in

  15. Simulation of a two-dimensional dipolar system on a APE100/quadrics SIMD architecture

    International Nuclear Information System (INIS)

    Bruno, A.; Pisacane, F.; Rosato, V.

    1997-01-01

    The temperature behavior of a system of dipoles with long-range interactions has been simulated via a two-dimensional lattice Monte Carlo on a massively (SIMD) platform (Quadrics/APE100). Thermodynamic quantities have been evaluated in order to locate and to characterize the phase transition in absence of applied field. Emphasis is given to the code implementation on the SIMD architecture and to the relevant features which have been used to improve code capabilities and performances. The probability of simultaneous occurrence of at least k spanning clusters has been studied by Monte Carlo simulations on the 2D square lattice with free boundaries at the bond percolation threshold p c = 1/2. It is found that the probability of k and more Incipient Spanning Clusters (ISC) have the values P(k > 1) ∼ 0.00658(3) and P(k > 2) ∼ 0.00000148(21) provided that the limit of these probabilities for infinite lattices exists. The probability P(k > 3) of more than three ISC could be estimated to be of the order of 10 -11 and is beyond the possibility to compute such a value by nowadays computers. So, it is impossible to check in simulations the Aizenman law for the probabilities when k much-gt 1. We have detected a single sample with four ISC in a total number of about 1010 samples investigated. The probability of this single event is 1/10 for that number of samples. The influence of boundary conditions is discussed in the last section

  16. Stream invertebrate productivity linked to forest subsidies: 37 stream-years of reference and experimental data

    Science.gov (United States)

    J. Bruce Wallace; Susan L Eggert; Judy L. Meyer; Jackson R. Webster

    2015-01-01

    Riparian habitats provide detrital subsidies of varying quantities and qualities to recipient ecosystems. We used long-term data from three reference streams (covering 24 stream-years) and 13-year whole-stream organic matter manipulations to investigate the influence of terrestrial detrital quantity and quality on benthic invertebrate community structure, abundance,...

  17. Information-Theoretic Data Discarding for Dynamic Trees on Data Streams

    Directory of Open Access Journals (Sweden)

    Christoforos Anagnostopoulos

    2013-12-01

    Full Text Available Ubiquitous automated data collection at an unprecedented scale is making available streaming, real-time information flows in a wide variety of settings, transforming both science and industry. Learning algorithms deployed in such contexts often rely on single-pass inference, where the data history is never revisited. Learning may also need to be temporally adaptive to remain up-to-date against unforeseen changes in the data generating mechanism. Online Bayesian inference remains challenged by such transient, evolving data streams. Nonparametric modeling techniques can prove particularly ill-suited, as the complexity of the model is allowed to increase with the sample size. In this work, we take steps to overcome these challenges by porting information theoretic heuristics, such as exponential forgetting and active learning, into a fully Bayesian framework. We showcase our methods by augmenting a modern non-parametric modeling framework, dynamic trees, and illustrate its performance on a number of practical examples. The end product is a powerful streaming regression and classification tool, whose performance compares favorably to the state-of-the-art.

  18. Exploring the connectome: Petascale volume visualization of microscopy data streams

    KAUST Repository

    Beyer, Johanna; Hadwiger, Markus; Al-Awami, Ali K.; Jeong, Wonki; Kasthuri, Narayanan; Lichtman, Jeff W M D; Pfister, Hanspeter

    2013-01-01

    Recent advances in high-resolution microscopy let neuroscientists acquire neural-tissue volume data of extremely large sizes. However, the tremendous resolution and the high complexity of neural structures present big challenges to storage, processing, and visualization at interactive rates. A proposed system provides interactive exploration of petascale (petavoxel) volumes resulting from high-throughput electron microscopy data streams. The system can concurrently handle multiple volumes and can support the simultaneous visualization of high-resolution voxel segmentation data. Its visualization-driven design restricts most computations to a small subset of the data. It employs a multiresolution virtual-memory architecture for better scalability than previous approaches and for handling incomplete data. Researchers have employed it for a 1-teravoxel mouse cortex volume, of which several hundred axons and dendrites as well as synapses have been segmented and labeled. © 1981-2012 IEEE.

  19. Exploring the connectome: Petascale volume visualization of microscopy data streams

    KAUST Repository

    Beyer, Johanna

    2013-07-01

    Recent advances in high-resolution microscopy let neuroscientists acquire neural-tissue volume data of extremely large sizes. However, the tremendous resolution and the high complexity of neural structures present big challenges to storage, processing, and visualization at interactive rates. A proposed system provides interactive exploration of petascale (petavoxel) volumes resulting from high-throughput electron microscopy data streams. The system can concurrently handle multiple volumes and can support the simultaneous visualization of high-resolution voxel segmentation data. Its visualization-driven design restricts most computations to a small subset of the data. It employs a multiresolution virtual-memory architecture for better scalability than previous approaches and for handling incomplete data. Researchers have employed it for a 1-teravoxel mouse cortex volume, of which several hundred axons and dendrites as well as synapses have been segmented and labeled. © 1981-2012 IEEE.

  20. KDE-Track: An Efficient Dynamic Density Estimator for Data Streams

    KAUST Repository

    Qahtan, Abdulhakim Ali Ali; Wang, Suojin; Zhang, Xiangliang

    2016-01-01

    Recent developments in sensors, global positioning system devices and smart phones have increased the availability of spatiotemporal data streams. Developing models for mining such streams is challenged by the huge amount of data that cannot be stored in the memory, the high arrival speed and the dynamic changes in the data distribution. Density estimation is an important technique in stream mining for a wide variety of applications. The construction of kernel density estimators is well studied and documented. However, existing techniques are either expensive or inaccurate and unable to capture the changes in the data distribution. In this paper, we present a method called KDE-Track to estimate the density of spatiotemporal data streams. KDE-Track can efficiently estimate the density function with linear time complexity using interpolation on a kernel model, which is incrementally updated upon the arrival of new samples from the stream. We also propose an accurate and efficient method for selecting the bandwidth value for the kernel density estimator, which increases its accuracy significantly. Both theoretical analysis and experimental validation show that KDE-Track outperforms a set of baseline methods on the estimation accuracy and computing time of complex density structures in data streams.

  1. KDE-Track: An Efficient Dynamic Density Estimator for Data Streams

    KAUST Repository

    Qahtan, Abdulhakim Ali Ali

    2016-11-08

    Recent developments in sensors, global positioning system devices and smart phones have increased the availability of spatiotemporal data streams. Developing models for mining such streams is challenged by the huge amount of data that cannot be stored in the memory, the high arrival speed and the dynamic changes in the data distribution. Density estimation is an important technique in stream mining for a wide variety of applications. The construction of kernel density estimators is well studied and documented. However, existing techniques are either expensive or inaccurate and unable to capture the changes in the data distribution. In this paper, we present a method called KDE-Track to estimate the density of spatiotemporal data streams. KDE-Track can efficiently estimate the density function with linear time complexity using interpolation on a kernel model, which is incrementally updated upon the arrival of new samples from the stream. We also propose an accurate and efficient method for selecting the bandwidth value for the kernel density estimator, which increases its accuracy significantly. Both theoretical analysis and experimental validation show that KDE-Track outperforms a set of baseline methods on the estimation accuracy and computing time of complex density structures in data streams.

  2. Biotic interactions modify multiple-stressor effects on juvenile brown trout in an experimental stream food web.

    Science.gov (United States)

    Bruder, Andreas; Salis, Romana K; Jones, Peter E; Matthaei, Christoph D

    2017-09-01

    Agricultural land use results in multiple stressors affecting stream ecosystems. Flow reduction due to water abstraction, elevated levels of nutrients and chemical contaminants are common agricultural stressors worldwide. Concurrently, stream ecosystems are also increasingly affected by climate change. Interactions among multiple co-occurring stressors result in biological responses that cannot be predicted from single-stressor effects (i.e. synergisms and antagonisms). At the ecosystem level, multiple-stressor effects can be further modified by biotic interactions (e.g. trophic interactions). We conducted a field experiment using 128 flow-through stream mesocosms to examine the individual and combined effects of water abstraction, nutrient enrichment and elevated levels of the nitrification inhibitor dicyandiamide (DCD) on survival, condition and gut content of juvenile brown trout and on benthic abundance of their invertebrate prey. Flow velocity reduction decreased fish survival (-12% compared to controls) and condition (-8% compared to initial condition), whereas effects of nutrient and DCD additions and interactions among these stressors were not significant. Negative effects of flow velocity reduction on fish survival and condition were consistent with effects on fish gut content (-25% compared to controls) and abundance of dominant invertebrate prey (-30% compared to controls), suggesting a negative metabolic balance driving fish mortality and condition decline, which was confirmed by structural equation modelling. Fish mortality under reduced flow velocity increased as maximal daily water temperatures approached the upper limit of their tolerance range, reflecting synergistic interactions between these stressors. Our study highlights the importance of indirect stressor effects such as those transferred through trophic interactions, which need to be considered when assessing and managing fish populations and stream food webs in multiple-stressor situations

  3. A Streaming PCA VLSI Chip for Neural Data Compression.

    Science.gov (United States)

    Wu, Tong; Zhao, Wenfeng; Guo, Hongsun; Lim, Hubert H; Yang, Zhi

    2017-12-01

    Neural recording system miniaturization and integration with low-power wireless technologies require compressing neural data before transmission. Feature extraction is a procedure to represent data in a low-dimensional space; its integration into a recording chip can be an efficient approach to compress neural data. In this paper, we propose a streaming principal component analysis algorithm and its microchip implementation to compress multichannel local field potential (LFP) and spike data. The circuits have been designed in a 65-nm CMOS technology and occupy a silicon area of 0.06 mm. Throughout the experiments, the chip compresses LFPs by 10 at the expense of as low as 1% reconstruction errors and 144-nW/channel power consumption; for spikes, the achieved compression ratio is 25 with 8% reconstruction errors and 3.05-W/channel power consumption. In addition, the algorithm and its hardware architecture can swiftly adapt to nonstationary spiking activities, which enables efficient hardware sharing among multiple channels to support a high-channel count recorder.

  4. Change Semantic Constrained Online Data Cleaning Method for Real-Time Observational Data Stream

    Science.gov (United States)

    Ding, Yulin; Lin, Hui; Li, Rongrong

    2016-06-01

    Recent breakthroughs in sensor networks have made it possible to collect and assemble increasing amounts of real-time observational data by observing dynamic phenomena at previously impossible time and space scales. Real-time observational data streams present potentially profound opportunities for real-time applications in disaster mitigation and emergency response, by providing accurate and timeliness estimates of environment's status. However, the data are always subject to inevitable anomalies (including errors and anomalous changes/events) caused by various effects produced by the environment they are monitoring. The "big but dirty" real-time observational data streams can rarely achieve their full potential in the following real-time models or applications due to the low data quality. Therefore, timely and meaningful online data cleaning is a necessary pre-requisite step to ensure the quality, reliability, and timeliness of the real-time observational data. In general, a straightforward streaming data cleaning approach, is to define various types of models/classifiers representing normal behavior of sensor data streams and then declare any deviation from this model as normal or erroneous data. The effectiveness of these models is affected by dynamic changes of deployed environments. Due to the changing nature of the complicated process being observed, real-time observational data is characterized by diversity and dynamic, showing a typical Big (Geo) Data characters. Dynamics and diversity is not only reflected in the data values, but also reflected in the complicated changing patterns of the data distributions. This means the pattern of the real-time observational data distribution is not stationary or static but changing and dynamic. After the data pattern changed, it is necessary to adapt the model over time to cope with the changing patterns of real-time data streams. Otherwise, the model will not fit the following observational data streams, which may led

  5. CHANGE SEMANTIC CONSTRAINED ONLINE DATA CLEANING METHOD FOR REAL-TIME OBSERVATIONAL DATA STREAM

    Directory of Open Access Journals (Sweden)

    Y. Ding

    2016-06-01

    Full Text Available Recent breakthroughs in sensor networks have made it possible to collect and assemble increasing amounts of real-time observational data by observing dynamic phenomena at previously impossible time and space scales. Real-time observational data streams present potentially profound opportunities for real-time applications in disaster mitigation and emergency response, by providing accurate and timeliness estimates of environment’s status. However, the data are always subject to inevitable anomalies (including errors and anomalous changes/events caused by various effects produced by the environment they are monitoring. The “big but dirty” real-time observational data streams can rarely achieve their full potential in the following real-time models or applications due to the low data quality. Therefore, timely and meaningful online data cleaning is a necessary pre-requisite step to ensure the quality, reliability, and timeliness of the real-time observational data. In general, a straightforward streaming data cleaning approach, is to define various types of models/classifiers representing normal behavior of sensor data streams and then declare any deviation from this model as normal or erroneous data. The effectiveness of these models is affected by dynamic changes of deployed environments. Due to the changing nature of the complicated process being observed, real-time observational data is characterized by diversity and dynamic, showing a typical Big (Geo Data characters. Dynamics and diversity is not only reflected in the data values, but also reflected in the complicated changing patterns of the data distributions. This means the pattern of the real-time observational data distribution is not stationary or static but changing and dynamic. After the data pattern changed, it is necessary to adapt the model over time to cope with the changing patterns of real-time data streams. Otherwise, the model will not fit the following observational

  6. Applications of spatial statistical network models to stream data

    Science.gov (United States)

    Daniel J. Isaak; Erin E. Peterson; Jay M. Ver Hoef; Seth J. Wenger; Jeffrey A. Falke; Christian E. Torgersen; Colin Sowder; E. Ashley Steel; Marie-Josee Fortin; Chris E. Jordan; Aaron S. Ruesch; Nicholas Som; Pascal. Monestiez

    2014-01-01

    Streams and rivers host a significant portion of Earth's biodiversity and provide important ecosystem services for human populations. Accurate information regarding the status and trends of stream resources is vital for their effective conservation and management. Most statistical techniques applied to data measured on stream networks were developed for...

  7. A framework to preserve the privacy of electronic health data streams.

    Science.gov (United States)

    Kim, Soohyung; Sung, Min Kyoung; Chung, Yon Dohn

    2014-08-01

    The anonymization of health data streams is important to protect these data against potential privacy breaches. A large number of research studies aiming at offering privacy in the context of data streams has been recently conducted. However, the techniques that have been proposed in these studies generate a significant delay during the anonymization process, since they concentrate on applying existing privacy models (e.g., k-anonymity and l-diversity) to batches of data extracted from data streams in a period of time. In this paper, we present delay-free anonymization, a framework for preserving the privacy of electronic health data streams. Unlike existing works, our method does not generate an accumulation delay, since input streams are anonymized immediately with counterfeit values. We further devise late validation for increasing the data utility of the anonymization results and managing the counterfeit values. Through experiments, we show the efficiency and effectiveness of the proposed method for the real-time release of data streams. Copyright © 2014 Elsevier Inc. All rights reserved.

  8. A SWAT model validation of nested-scale contemporaneous stream flow, suspended sediment and nutrients from a multiple-land-use watershed of the central USA.

    Science.gov (United States)

    Zeiger, Sean J; Hubbart, Jason A

    2016-12-01

    There is an ongoing need to validate the accuracy of predictive model simulated pollutant yields, particularly from multiple-land-use (i.e. forested, agricultural, and urban) watersheds. However, there are seldom sufficient observed data sets available that supply requisite spatial and temporal resolution and coupled multi-parameter constituents for rigorous model performance assessment. Four years of hydroclimate and water quality data were used to validate SWAT model estimates of monthly stream flow, suspended sediment, total phosphorus, nitrate, nitrite, ammonium, and total inorganic nitrogen from 5 nested-scale gauging sites located in a multiple-land-use watershed of the central USA. The uncalibrated SWAT model satisfactorily simulated monthly stream flow with Nash-Sutcliffe efficiency (NSE) values ranging from 0.50 near the headwaters, to 0.75 near the watershed outlet. However, the uncalibrated model did not accurately simulate monthly sediment, total phosphorus, nitrate, nitrite, ammonium, and total inorganic nitrogen with NSE valuesSWAT model to multiple gauging sites within the watershed improved estimates of monthly stream flow (NSE=0.83), sediment (NSE=0.78), total phosphorus (NSE=0.81), nitrate (NSE=0.90), and total inorganic nitrogen (NSE=0.86). However, NSE values were model performance decreased for sediment, nitrate, and total inorganic nitrogen during the validation period with NSE valuesSWAT model to multiple gauging sites and provide guidance to SWAT model (or similar models) users wishing to improve model performance at multiple scales. Copyright © 2016 Elsevier B.V. All rights reserved.

  9. Fast algorithm for automatically computing Strahler stream order

    Science.gov (United States)

    Lanfear, Kenneth J.

    1990-01-01

    An efficient algorithm was developed to determine Strahler stream order for segments of stream networks represented in a Geographic Information System (GIS). The algorithm correctly assigns Strahler stream order in topologically complex situations such as braided streams and multiple drainage outlets. Execution time varies nearly linearly with the number of stream segments in the network. This technique is expected to be particularly useful for studying the topology of dense stream networks derived from digital elevation model data.

  10. Elastic execution of continuous mapreduce jobs over data streams

    DEFF Research Database (Denmark)

    2015-01-01

    There is provided a set of methods describing how to elastically change the resources used by a MapReduce job on streaming data while executing......There is provided a set of methods describing how to elastically change the resources used by a MapReduce job on streaming data while executing...

  11. Efficient Estimation of Dynamic Density Functions with Applications in Streaming Data

    KAUST Repository

    Qahtan, Abdulhakim

    2016-05-11

    Recent advances in computing technology allow for collecting vast amount of data that arrive continuously in the form of streams. Mining data streams is challenged by the speed and volume of the arriving data. Furthermore, the underlying distribution of the data changes over the time in unpredicted scenarios. To reduce the computational cost, data streams are often studied in forms of condensed representation, e.g., Probability Density Function (PDF). This thesis aims at developing an online density estimator that builds a model called KDE-Track for characterizing the dynamic density of the data streams. KDE-Track estimates the PDF of the stream at a set of resampling points and uses interpolation to estimate the density at any given point. To reduce the interpolation error and computational complexity, we introduce adaptive resampling where more/less resampling points are used in high/low curved regions of the PDF. The PDF values at the resampling points are updated online to provide up-to-date model of the data stream. Comparing with other existing online density estimators, KDE-Track is often more accurate (as reflected by smaller error values) and more computationally efficient (as reflected by shorter running time). The anytime available PDF estimated by KDE-Track can be applied for visualizing the dynamic density of data streams, outlier detection and change detection in data streams. In this thesis work, the first application is to visualize the taxi traffic volume in New York city. Utilizing KDE-Track allows for visualizing and monitoring the traffic flow on real time without extra overhead and provides insight analysis of the pick up demand that can be utilized by service providers to improve service availability. The second application is to detect outliers in data streams from sensor networks based on the estimated PDF. The method detects outliers accurately and outperforms baseline methods designed for detecting and cleaning outliers in sensor data. The

  12. Stream invertebrate productivity linked to forest subsidies: 37 stream-years of reference and experimental data.

    Science.gov (United States)

    Wallace, J Bruce; Eggert, Susan L; Meyer, Judy L; Webster, Jackson R

    2015-05-01

    Riparian habitats provide detrital subsidies of varying quantities and qualities to recipient ecosystems. We used long-term data from three reference streams (covering 24 stream-years) and 13-year whole-stream organic matter manipulations to investigate the influence of terrestrial detrital quantity and quality on benthic invertebrate community structure, abundance, biomass, and secondary production in rockface (RF) and mixed substrates (MS) of forested headwater streams. Using a mesh canopy covering the entire treatment stream, we examined effects of litter ex'clusion, small- and large-wood removal, and addition of artificial wood (PVC) and leaves of varying quality on organic matter standing crops and invertebrate community structure and function. We assessed differences in functional feeding group distribution between substrate types as influenced by organic matter manipulations and long-term patterns of predator and prey production in manipulated vs. reference years. Particulate organic matter standing crops in MS of the treatment stream declined drastically with each successive year of litter exclusion, approaching zero after three years. Monthly invertebrate biomass and annual secondary production was positively related to benthic organic matter in the MS habitats. Rockface habitats exhibited fewer changes than MS habitats across all organic matter manipulations. With leaf addition, the patterns of functional group distribution among MS and RF habitats returned to patterns seen in reference streams. Secondary production per unit organic matter standing crop was greatest for the leaf addition period, followed by the reference streams, and significantly less for the litter exclusion and wood removal periods. These data indicate that the limited organic matter remaining in the stream following litter exclusion and wood removal was more refractory than that in the reference streams, whereas the added leaf material was more labile and readily converted into

  13. Developing a national stream morphology data exchange: Needs, challenges, and opportunities

    Science.gov (United States)

    Collins, Mathias J.; Gray, John R.; Peppler, Marie C.; Fitzpatrick, Faith A.; Schubauer-Berigan, Joseph P.

    2012-05-01

    Stream morphology data, primarily consisting of channel and foodplain geometry and bed material size measurements, historically have had a wide range of applications and uses including culvert/ bridge design, rainfall- runoff modeling, food inundation mapping (e.g., U.S. Federal Emergency Management Agency food insurance studies), climate change studies, channel stability/sediment source investigations, navigation studies, habitat assessments, and landscape change research. The need for stream morphology data in the United States, and thus the quantity of data collected, has grown substantially over the past 2 decades because of the expanded interests of resource management agencies in watershed management and restoration. The quantity of stream morphology data collected has also increased because of state-of-the-art technologies capable of rapidly collecting high-resolution data over large areas with heretofore unprecedented precision. Despite increasing needs for and the expanding quantity of stream morphology data, neither common reporting standards nor a central data archive exist for storing and serving these often large and spatially complex data sets. We are proposing an open- access data exchange for archiving and disseminating stream morphology data.

  14. Monitoring forest cover loss using multiple data streams, a case study of a tropical dry forest in Bolivia

    Science.gov (United States)

    Dutrieux, Loïc Paul; Verbesselt, Jan; Kooistra, Lammert; Herold, Martin

    2015-09-01

    Automatically detecting forest disturbances as they occur can be extremely challenging for certain types of environments, particularly those presenting strong natural variations. Here, we use a generic structural break detection framework (BFAST) to improve the monitoring of forest cover loss by combining multiple data streams. Forest change monitoring is performed using Landsat data in combination with MODIS or rainfall data to further improve the modelling and monitoring. We tested the use of the Normalized Difference Vegetation Index (NDVI) from the Moderate Resolution Imaging Spectroradiometer (MODIS) with varying spatial aggregation window sizes as well as a rainfall derived index as external regressors. The method was evaluated on a dry tropical forest area in lowland Bolivia where forest cover loss is known to occur, and we validated the results against a set of ground truth samples manually interpreted using the TimeSync environment. We found that the addition of an external regressor allows to take advantage of the difference in spatial extent between human induced and naturally induced variations and only detect the processes of interest. Of all configurations, we found the 13 by 13 km MODIS NDVI window to be the most successful, with an overall accuracy of 87%. Compared with a single pixel approach, the proposed method produced better time-series model fits resulting in increases of overall accuracy (from 82% to 87%), and decrease in omission and commission errors (from 33% to 24% and from 3% to 0% respectively). The presented approach seems particularly relevant for areas with high inter-annual natural variability, such as forests regularly experiencing exceptional drought events.

  15. Frequent Pairs in Data Streams: Exploiting Parallelism and Skew

    DEFF Research Database (Denmark)

    Campagna, Andrea; Kutzkow, Konstantin; Pagh, Rasmus

    2011-01-01

    We introduce the Pair Streaming Engine (PairSE) that detects frequent pairs in a data stream of transactions. Our algorithm finds the most frequent pairs with high probability, and gives tight bounds on their frequency. It is particularly space efficient for skewed distribution of pair supports...... items mining in data streams. We show how to efficiently scale these approaches to handle large transactions. We report experimental results showcasing precision and recall of our method. In particular, we find that often our method achieves excellent precision, returning identical upper and lower...... bounds on the supports of the most frequent pairs....

  16. Self-adaptive change detection in streaming data with non-stationary distribution

    KAUST Repository

    Zhang, Xiangliang

    2010-01-01

    Non-stationary distribution, in which the data distribution evolves over time, is a common issue in many application fields, e.g., intrusion detection and grid computing. Detecting the changes in massive streaming data with a non-stationary distribution helps to alarm the anomalies, to clean the noises, and to report the new patterns. In this paper, we employ a novel approach for detecting changes in streaming data with the purpose of improving the quality of modeling the data streams. Through observing the outliers, this approach of change detection uses a weighted standard deviation to monitor the evolution of the distribution of data streams. A cumulative statistical test, Page-Hinkley, is employed to collect the evidence of changes in distribution. The parameter used for reporting the changes is self-adaptively adjusted according to the distribution of data streams, rather than set by a fixed empirical value. The self-adaptability of the novel approach enhances the effectiveness of modeling data streams by timely catching the changes of distributions. We validated the approach on an online clustering framework with a benchmark KDDcup 1999 intrusion detection data set as well as with a real-world grid data set. The validation results demonstrate its better performance on achieving higher accuracy and lower percentage of outliers comparing to the other change detection approaches. © 2010 Springer-Verlag.

  17. Analysis of sound data streamed over the network

    Directory of Open Access Journals (Sweden)

    Jiří Fejfar

    2013-01-01

    Full Text Available In this paper we inspect a difference between original sound recording and signal captured after streaming this original recording over a network loaded with a heavy traffic. There are several kinds of failures occurring in the captured recording caused by network congestion. We try to find a method how to evaluate correctness of streamed audio. Usually there are metrics based on a human perception of a signal such as “signal is clear, without audible failures”, “signal is having some failures but it is understandable”, or “signal is inarticulate”. These approaches need to be statistically evaluated on a broad set of respondents, which is time and resource consuming. We try to propose some metrics based on signal properties allowing us to compare the original and captured recording. We use algorithm called Dynamic Time Warping (Müller, 2007 commonly used for time series comparison in this paper. Some other time series exploration approaches can be found in (Fejfar, 2011 and (Fejfar, 2012. The data was acquired in our network laboratory simulating network traffic by downloading files, streaming audio and video simultaneously. Our former experiment inspected Quality of Service (QoS and its impact on failures of received audio data stream. This experiment is focused on the comparison of sound recordings rather than network mechanism.We focus, in this paper, on a real time audio stream such as a telephone call, where it is not possible to stream audio in advance to a “pool”. Instead it is necessary to achieve as small delay as possible (between speaker voice recording and listener voice replay. We are using RTP protocol for streaming audio.

  18. A Quality-Centric Data Model for Distributed Stream Management Systems

    OpenAIRE

    Pietzuch, P; Fiscato, M; Vu, QH

    2009-01-01

    21.10.14 KB Ok to add published version to spiral. It is challenging for large-scale stream management systems to return always perfect results when processing data streams originating from distributed sources. Data sources and intermediate processing nodes may fail during the lifetime of a stream query. In addition, individual nodes may become overloaded due to processing demands. In practice, users have to accept incomplete or inaccurate query results because of failure or overload. In t...

  19. Ten years of multiple data stream assimilation with the ORCHIDEE land surface model to improve regional to global simulated carbon budgets: synthesis and perspectives on directions for the future

    Science.gov (United States)

    Peylin, P. P.; Bacour, C.; MacBean, N.; Maignan, F.; Bastrikov, V.; Chevallier, F.

    2017-12-01

    Predicting the fate of carbon stocks and their sensitivity to climate change and land use/management strongly relies on our ability to accurately model net and gross carbon fluxes. However, simulated carbon and water fluxes remain subject to large uncertainties, partly because of unknown or poorly calibrated parameters. Over the past ten years, the carbon cycle data assimilation system at the Laboratoire des Sciences du Climat et de l'Environnement has investigated the benefit of assimilating multiple carbon cycle data streams into the ORCHIDEE LSM, the land surface component of the Institut Pierre Simon Laplace Earth System Model. These datasets have included FLUXNET eddy covariance data (net CO2 flux and latent heat flux) to constrain hourly to seasonal time-scale carbon cycle processes, remote sensing of the vegetation activity (MODIS NDVI) to constrain the leaf phenology, biomass data to constrain "slow" (yearly to decadal) processes of carbon allocation, and atmospheric CO2 concentrations to provide overall large scale constraints on the land carbon sink. Furthermore, we have investigated technical issues related to multiple data stream assimilation and choice of optimization algorithm. This has provided a wide-ranging perspective on the challenges we face in constraining model parameters and thus better quantifying, and reducing, model uncertainty in projections of the future global carbon sink. We review our past studies in terms of the impact of the optimization on key characteristics of the carbon cycle, e.g. the partition of the northern latitudes vs tropical land carbon sink, and compare to the classic atmospheric flux inversion approach. Throughout, we discuss our work in context of the abovementioned challenges, and propose solutions for the community going forward, including the potential of new observations such as atmospheric COS concentrations and satellite-derived Solar Induced Fluorescence to constrain the gross carbon fluxes of the ORCHIDEE

  20. Clustering big data streams : recent challenges and contributions

    NARCIS (Netherlands)

    Hassani, M.; Seidl, T.

    Traditional clustering algorithms merely considered static data. Today's various applications and research issues in big data mining have however to deal with continuous, possibly infinite streams of data, arriving at high velocity. Web traffic data, surveillance data, sensor measurements and stock

  1. CUDASW++2.0: enhanced Smith-Waterman protein database search on CUDA-enabled GPUs based on SIMT and virtualized SIMD abstractions

    Directory of Open Access Journals (Sweden)

    Schmidt Bertil

    2010-04-01

    Full Text Available Abstract Background Due to its high sensitivity, the Smith-Waterman algorithm is widely used for biological database searches. Unfortunately, the quadratic time complexity of this algorithm makes it highly time-consuming. The exponential growth of biological databases further deteriorates the situation. To accelerate this algorithm, many efforts have been made to develop techniques in high performance architectures, especially the recently emerging many-core architectures and their associated programming models. Findings This paper describes the latest release of the CUDASW++ software, CUDASW++ 2.0, which makes new contributions to Smith-Waterman protein database searches using compute unified device architecture (CUDA. A parallel Smith-Waterman algorithm is proposed to further optimize the performance of CUDASW++ 1.0 based on the single instruction, multiple thread (SIMT abstraction. For the first time, we have investigated a partitioned vectorized Smith-Waterman algorithm using CUDA based on the virtualized single instruction, multiple data (SIMD abstraction. The optimized SIMT and the partitioned vectorized algorithms were benchmarked, and remarkably, have similar performance characteristics. CUDASW++ 2.0 achieves performance improvement over CUDASW++ 1.0 as much as 1.74 (1.72 times using the optimized SIMT algorithm and up to 1.77 (1.66 times using the partitioned vectorized algorithm, with a performance of up to 17 (30 billion cells update per second (GCUPS on a single-GPU GeForce GTX 280 (dual-GPU GeForce GTX 295 graphics card. Conclusions CUDASW++ 2.0 is publicly available open-source software, written in CUDA and C++ programming languages. It obtains significant performance improvement over CUDASW++ 1.0 using either the optimized SIMT algorithm or the partitioned vectorized algorithm for Smith-Waterman protein database searches by fully exploiting the compute capability of commonly used CUDA-enabled low-cost GPUs.

  2. Stream Kriging: Incremental and recursive ordinary Kriging over spatiotemporal data streams

    Science.gov (United States)

    Zhong, Xu; Kealy, Allison; Duckham, Matt

    2016-05-01

    Ordinary Kriging is widely used for geospatial interpolation and estimation. Due to the O (n3) time complexity of solving the system of linear equations, ordinary Kriging for a large set of source points is computationally intensive. Conducting real-time Kriging interpolation over continuously varying spatiotemporal data streams can therefore be especially challenging. This paper develops and tests two new strategies for improving the performance of an ordinary Kriging interpolator adapted to a stream-processing environment. These strategies rely on the expectation that, over time, source data points will frequently refer to the same spatial locations (for example, where static sensor nodes are generating repeated observations of a dynamic field). First, an incremental strategy improves efficiency in cases where a relatively small proportion of previously processed spatial locations are absent from the source points at any given iteration. Second, a recursive strategy improves efficiency in cases where there is substantial set overlap between the sets of spatial locations of source points at the current and previous iterations. These two strategies are evaluated in terms of their computational efficiency in comparison to ordinary Kriging algorithm. The results show that these two strategies can reduce the time taken to perform the interpolation by up to 90%, and approach average-case time complexity of O (n2) when most but not all source points refer to the same locations over time. By combining the approaches developed in this paper with existing heuristic ordinary Kriging algorithms, the conclusions indicate how further efficiency gains could potentially be accrued. The work ultimately contributes to the development of online ordinary Kriging interpolation algorithms, capable of real-time spatial interpolation with large streaming data sets.

  3. rEMM: Extensible Markov Model for Data Stream Clustering in R

    Directory of Open Access Journals (Sweden)

    Michael Hahsler

    2010-10-01

    Full Text Available Clustering streams of continuously arriving data has become an important application of data mining in recent years and efficient algorithms have been proposed by several researchers. However, clustering alone neglects the fact that data in a data stream is not only characterized by the proximity of data points which is used by clustering, but also by a temporal component. The extensible Markov model (EMM adds the temporal component to data stream clustering by superimposing a dynamically adapting Markov chain. In this paper we introduce the implementation of the R extension package rEMM which implements EMM and we discuss some examples and applications.

  4. Sampling the stream landscape: Improving the applicability of an ecoregion-level capture probability model for stream fishes

    Science.gov (United States)

    Mollenhauer, Robert; Mouser, Joshua B.; Brewer, Shannon K.

    2018-01-01

    Temporal and spatial variability in streams result in heterogeneous gear capture probability (i.e., the proportion of available individuals identified) that confounds interpretation of data used to monitor fish abundance. We modeled tow-barge electrofishing capture probability at multiple spatial scales for nine Ozark Highland stream fishes. In addition to fish size, we identified seven reach-scale environmental characteristics associated with variable capture probability: stream discharge, water depth, conductivity, water clarity, emergent vegetation, wetted width–depth ratio, and proportion of riffle habitat. The magnitude of the relationship between capture probability and both discharge and depth varied among stream fishes. We also identified lithological characteristics among stream segments as a coarse-scale source of variable capture probability. The resulting capture probability model can be used to adjust catch data and derive reach-scale absolute abundance estimates across a wide range of sampling conditions with similar effort as used in more traditional fisheries surveys (i.e., catch per unit effort). Adjusting catch data based on variable capture probability improves the comparability of data sets, thus promoting both well-informed conservation and management decisions and advances in stream-fish ecology.

  5. A data-stream classification system for investigating terrorist threats

    Science.gov (United States)

    Schulz, Alexia; Dettman, Joshua; Gottschalk, Jeffrey; Kotson, Michael; Vuksani, Era; Yu, Tamara

    2016-05-01

    The role of cyber forensics in criminal investigations has greatly increased in recent years due to the wealth of data that is collected and available to investigators. Physical forensics has also experienced a data volume and fidelity revolution due to advances in methods for DNA and trace evidence analysis. Key to extracting insight is the ability to correlate across multi-modal data, which depends critically on identifying a touch-point connecting the separate data streams. Separate data sources may be connected because they refer to the same individual, entity or event. In this paper we present a data source classification system tailored to facilitate the investigation of potential terrorist activity. This taxonomy is structured to illuminate the defining characteristics of a particular terrorist effort and designed to guide reporting to decision makers that is complete, concise, and evidence-based. The classification system has been validated and empirically utilized in the forensic analysis of a simulated terrorist activity. Next-generation analysts can use this schema to label and correlate across existing data streams, assess which critical information may be missing from the data, and identify options for collecting additional data streams to fill information gaps.

  6. New Splitting Criteria for Decision Trees in Stationary Data Streams.

    Science.gov (United States)

    Jaworski, Maciej; Duda, Piotr; Rutkowski, Leszek; Jaworski, Maciej; Duda, Piotr; Rutkowski, Leszek; Rutkowski, Leszek; Duda, Piotr; Jaworski, Maciej

    2018-06-01

    The most popular tools for stream data mining are based on decision trees. In previous 15 years, all designed methods, headed by the very fast decision tree algorithm, relayed on Hoeffding's inequality and hundreds of researchers followed this scheme. Recently, we have demonstrated that although the Hoeffding decision trees are an effective tool for dealing with stream data, they are a purely heuristic procedure; for example, classical decision trees such as ID3 or CART cannot be adopted to data stream mining using Hoeffding's inequality. Therefore, there is an urgent need to develop new algorithms, which are both mathematically justified and characterized by good performance. In this paper, we address this problem by developing a family of new splitting criteria for classification in stationary data streams and investigating their probabilistic properties. The new criteria, derived using appropriate statistical tools, are based on the misclassification error and the Gini index impurity measures. The general division of splitting criteria into two types is proposed. Attributes chosen based on type- splitting criteria guarantee, with high probability, the highest expected value of split measure. Type- criteria ensure that the chosen attribute is the same, with high probability, as it would be chosen based on the whole infinite data stream. Moreover, in this paper, two hybrid splitting criteria are proposed, which are the combinations of single criteria based on the misclassification error and Gini index.

  7. An efficient reversible privacy-preserving data mining technology over data streams.

    Science.gov (United States)

    Lin, Chen-Yi; Kao, Yuan-Hung; Lee, Wei-Bin; Chen, Rong-Chang

    2016-01-01

    With the popularity of smart handheld devices and the emergence of cloud computing, users and companies can save various data, which may contain private data, to the cloud. Topics relating to data security have therefore received much attention. This study focuses on data stream environments and uses the concept of a sliding window to design a reversible privacy-preserving technology to process continuous data in real time, known as a continuous reversible privacy-preserving (CRP) algorithm. Data with CRP algorithm protection can be accurately recovered through a data recovery process. In addition, by using an embedded watermark, the integrity of the data can be verified. The results from the experiments show that, compared to existing algorithms, CRP is better at preserving knowledge and is more effective in terms of reducing information loss and privacy disclosure risk. In addition, it takes far less time for CRP to process continuous data than existing algorithms. As a result, CRP is confirmed as suitable for data stream environments and fulfills the requirements of being lightweight and energy-efficient for smart handheld devices.

  8. Encrypted data stream identification using randomness sparse representation and fuzzy Gaussian mixture model

    Science.gov (United States)

    Zhang, Hong; Hou, Rui; Yi, Lei; Meng, Juan; Pan, Zhisong; Zhou, Yuhuan

    2016-07-01

    The accurate identification of encrypted data stream helps to regulate illegal data, detect network attacks and protect users' information. In this paper, a novel encrypted data stream identification algorithm is introduced. The proposed method is based on randomness characteristics of encrypted data stream. We use a l1-norm regularized logistic regression to improve sparse representation of randomness features and Fuzzy Gaussian Mixture Model (FGMM) to improve identification accuracy. Experimental results demonstrate that the method can be adopted as an effective technique for encrypted data stream identification.

  9. Learning from medical data streams: an introduction

    NARCIS (Netherlands)

    Pereira Rodrigues, P.; Pechenizkiy, M.; Gaber, M.M.; Gama, J.

    2011-01-01

    Clinical practice and research are facing a new challenge created by the rapid growth of health information science and technology, and the complexity and volume of biomedical data. Machine learning from medical data streams is a recent area of research that aims to provide better knowledge

  10. National Aquatic Resource Survey Rivers and Streams Data

    Data.gov (United States)

    U.S. Environmental Protection Agency — Data are from 1,000 river and stream sites across the conterminous US where consistent biological, chemical, physical and watershed data were gathered. The sites...

  11. Local correlation detection with linearity enhancement in streaming data

    KAUST Repository

    Xie, Qing

    2013-01-01

    This paper addresses the challenges in detecting the potential correlation between numerical data streams, which facilitates the research of data stream mining and pattern discovery. We focus on local correlation with delay, which may occur in burst at different time in different streams, and last for a limited period. The uncertainty on the correlation occurrence and the time delay make it diff cult to monitor the correlation online. Furthermore, the conventional correlation measure lacks the ability of ref ecting visual linearity, which is more desirable in reality. This paper proposes effective methods to continuously detect the correlation between data streams. Our approach is based on the Discrete Fourier Transform to make rapid cross-correlation calculation with time delay allowed. In addition, we introduce a shape-based similarity measure into the framework, which ref nes the results by representative trend patterns to enhance the signif cance of linearity. The similarity of proposed linear representations can quickly estimate the correlation, and the window sliding strategy in segment level improves the eff ciency for online detection. The empirical study demonstrates the accuracy of our detection approach, as well as more than 30% improvement of eff ciency. Copyright 2013 ACM.

  12. Real-Time Clinical Decision Support System with Data Stream Mining

    Directory of Open Access Journals (Sweden)

    Yang Zhang

    2012-01-01

    Full Text Available This research aims to describe a new design of data stream mining system that can analyze medical data stream and make real-time prediction. The motivation of the research is due to a growing concern of combining software technology and medical functions for the development of software application that can be used in medical field of chronic disease prognosis and diagnosis, children healthcare, diabetes diagnosis, and so forth. Most of the existing software technologies are case-based data mining systems. They only can analyze finite and structured data set and can only work well in their early years and can hardly meet today's medical requirement. In this paper, we describe a clinical-support-system based data stream mining technology; the design has taken into account all the shortcomings of the existing clinical support systems.

  13. Real-time analytics techniques to analyze and visualize streaming data

    CERN Document Server

    Ellis, Byron

    2014-01-01

    Construct a robust end-to-end solution for analyzing and visualizing streaming data Real-time analytics is the hottest topic in data analytics today. In Real-Time Analytics: Techniques to Analyze and Visualize Streaming Data, expert Byron Ellis teaches data analysts technologies to build an effective real-time analytics platform. This platform can then be used to make sense of the constantly changing data that is beginning to outpace traditional batch-based analysis platforms. The author is among a very few leading experts in the field. He has a prestigious background in research, development,

  14. Mutual Information Based Dynamic Integration of Multiple Feature Streams for Robust Real-Time LVCSR

    Science.gov (United States)

    Sato, Shoei; Kobayashi, Akio; Onoe, Kazuo; Homma, Shinichi; Imai, Toru; Takagi, Tohru; Kobayashi, Tetsunori

    We present a novel method of integrating the likelihoods of multiple feature streams, representing different acoustic aspects, for robust speech recognition. The integration algorithm dynamically calculates a frame-wise stream weight so that a higher weight is given to a stream that is robust to a variety of noisy environments or speaking styles. Such a robust stream is expected to show discriminative ability. A conventional method proposed for the recognition of spoken digits calculates the weights front the entropy of the whole set of HMM states. This paper extends the dynamic weighting to a real-time large-vocabulary continuous speech recognition (LVCSR) system. The proposed weight is calculated in real-time from mutual information between an input stream and active HMM states in a searchs pace without an additional likelihood calculation. Furthermore, the mutual information takes the width of the search space into account by calculating the marginal entropy from the number of active states. In this paper, we integrate three features that are extracted through auditory filters by taking into account the human auditory system's ability to extract amplitude and frequency modulations. Due to this, features representing energy, amplitude drift, and resonant frequency drifts, are integrated. These features are expected to provide complementary clues for speech recognition. Speech recognition experiments on field reports and spontaneous commentary from Japanese broadcast news showed that the proposed method reduced error words by 9.2% in field reports and 4.7% in spontaneous commentaries relative to the best result obtained from a single stream.

  15. A PCA-Based Change Detection Framework for Multidimensional Data Streams

    KAUST Repository

    Qahtan, Abdulhakim Ali Ali

    2015-08-10

    Detecting changes in multidimensional data streams is an important and challenging task. In unsupervised change detection, changes are usually detected by comparing the distribution in a current (test) window with a reference window. It is thus essential to design divergence metrics and density estimators for comparing the data distributions, which are mostly done for univariate data. Detecting changes in multidimensional data streams brings difficulties to the density estimation and comparisons. In this paper, we propose a framework for detecting changes in multidimensional data streams based on principal component analysis, which is used for projecting data into a lower dimensional space, thus facilitating density estimation and change-score calculations. The proposed framework also has advantages over existing approaches by reducing computational costs with an efficient density estimator, promoting the change-score calculation by introducing effective divergence metrics, and by minimizing the efforts required from users on the threshold parameter setting by using the Page-Hinkley test. The evaluation results on synthetic and real data show that our framework outperforms two baseline methods in terms of both detection accuracy and computational costs.

  16. FACT. Streamed data analysis and online application of machine learning models

    Energy Technology Data Exchange (ETDEWEB)

    Bruegge, Kai Arno; Buss, Jens [Technische Universitaet Dortmund (Germany). Astroteilchenphysik; Collaboration: FACT-Collaboration

    2016-07-01

    Imaging Atmospheric Cherenkov Telescopes (IACTs) like FACT produce a continuous flow of data during measurements. Analyzing the data in near real time is essential for monitoring sources. One major task of a monitoring system is to detect changes in the gamma-ray flux of a source, and to alert other experiments if some predefined limit is reached. In order to calculate the flux of an observed source, it is necessary to run an entire data analysis process including calibration, image cleaning, parameterization, signal-background separation and flux estimation. Software built on top of a data streaming framework has been implemented for FACT and generalized to work with the data acquisition framework of the Cherenkov Telescope Array (CTA). We present how the streams-framework is used to apply supervised machine learning models to an online data stream from the telescope.

  17. Object tracking using multiple camera video streams

    Science.gov (United States)

    Mehrubeoglu, Mehrube; Rojas, Diego; McLauchlan, Lifford

    2010-05-01

    Two synchronized cameras are utilized to obtain independent video streams to detect moving objects from two different viewing angles. The video frames are directly correlated in time. Moving objects in image frames from the two cameras are identified and tagged for tracking. One advantage of such a system involves overcoming effects of occlusions that could result in an object in partial or full view in one camera, when the same object is fully visible in another camera. Object registration is achieved by determining the location of common features in the moving object across simultaneous frames. Perspective differences are adjusted. Combining information from images from multiple cameras increases robustness of the tracking process. Motion tracking is achieved by determining anomalies caused by the objects' movement across frames in time in each and the combined video information. The path of each object is determined heuristically. Accuracy of detection is dependent on the speed of the object as well as variations in direction of motion. Fast cameras increase accuracy but limit the speed and complexity of the algorithm. Such an imaging system has applications in traffic analysis, surveillance and security, as well as object modeling from multi-view images. The system can easily be expanded by increasing the number of cameras such that there is an overlap between the scenes from at least two cameras in proximity. An object can then be tracked long distances or across multiple cameras continuously, applicable, for example, in wireless sensor networks for surveillance or navigation.

  18. Fault-Tolerant and Elastic Streaming MapReduce with Decentralized Coordination

    Energy Technology Data Exchange (ETDEWEB)

    Kumbhare, Alok [Univ. of Southern California, Los Angeles, CA (United States); Frincu, Marc [Univ. of Southern California, Los Angeles, CA (United States); Simmhan, Yogesh [Indian Inst. of Technology (IIT), Bangalore (India); Prasanna, Viktor K. [Univ. of Southern California, Los Angeles, CA (United States)

    2015-06-29

    The MapReduce programming model, due to its simplicity and scalability, has become an essential tool for processing large data volumes in distributed environments. Recent Stream Processing Systems (SPS) extend this model to provide low-latency analysis of high-velocity continuous data streams. However, integrating MapReduce with streaming poses challenges: first, the runtime variations in data characteristics such as data-rates and key-distribution cause resource overload, that inturn leads to fluctuations in the Quality of the Service (QoS); and second, the stateful reducers, whose state depends on the complete tuple history, necessitates efficient fault-recovery mechanisms to maintain the desired QoS in the presence of resource failures. We propose an integrated streaming MapReduce architecture leveraging the concept of consistent hashing to support runtime elasticity along with locality-aware data and state replication to provide efficient load-balancing with low-overhead fault-tolerance and parallel fault-recovery from multiple simultaneous failures. Our evaluation on a private cloud shows up to 2:8 improvement in peak throughput compared to Apache Storm SPS, and a low recovery latency of 700 -1500 ms from multiple failures.

  19. A Stream Tilling Approach to Surface Area Estimation for Large Scale Spatial Data in a Shared Memory System

    Directory of Open Access Journals (Sweden)

    Liu Jiping

    2017-12-01

    Full Text Available Surface area estimation is a widely used tool for resource evaluation in the physical world. When processing large scale spatial data, the input/output (I/O can easily become the bottleneck in parallelizing the algorithm due to the limited physical memory resources and the very slow disk transfer rate. In this paper, we proposed a stream tilling approach to surface area estimation that first decomposed a spatial data set into tiles with topological expansions. With these tiles, the one-to-one mapping relationship between the input and the computing process was broken. Then, we realized a streaming framework towards the scheduling of the I/O processes and computing units. Herein, each computing unit encapsulated a same copy of the estimation algorithm, and multiple asynchronous computing units could work individually in parallel. Finally, the performed experiment demonstrated that our stream tilling estimation can efficiently alleviate the heavy pressures from the I/O-bound work, and the measured speedup after being optimized have greatly outperformed the directly parallel versions in shared memory systems with multi-core processors.

  20. StreamNet, 1999-2000 Annual Report.

    Energy Technology Data Exchange (ETDEWEB)

    Schmidt, Bruce; Roger, Phil; Butterfield, Bart (Pacific States Marine Fisheries Commission, Gladstone, OR)

    2001-09-01

    The StreamNet Project is a cooperative project that provides basic fishery management data in a consistent format across the Columbia Basin region, with some data from outside the region. Specific categories of data are acquired from the multiple data generating agencies in the Columbia Basin, converted into a standardized data exchange format (DEF) and distributed to fish researchers, managers and decision makers directly or through an on-line data retrieval system (www.streamnet.org). The project is funded by the Bonneville Power Administration (BPA) as part of the Northwest Power Planning Council's (NWPPC) Fish and Wildlife Program. This cooperative effort is composed of a region-wide project administered by the Pacific States Marine Fisheries Commission (PSMFC) that is responsible for project management, regional data management and data delivery (Region), plus seven contributing projects within the data generating entities: Columbia River Intertribal Fish Commission (CRITFC); Idaho Department of Fish and Game (IDFG); Montana Fish, Wildlife and Parks (MFWP); Oregon Department of Fish and Wildlife (ODFW); Shoshone-Bannock Tribes; U. S. Fish and Wildlife Service (FWS); and Washington Department of Fish and Wildlife (WDFW). The contributing projects are funded through the StreamNet contract but work within their respective agencies and are referred to here as the agency's StreamNet project (for example, ''IDFG StreamNet'' for Idaho's project). The StreamNet Project provides an important link in the chain of data flow in the Columbia Basin, with specific emphasis on data collected routinely over time by management agencies. Basic fish related data are collected in the field by the various state, tribal and federal agencies in the basin for purposes related to each agency's individual mission and responsibility. As a result, there often is a lack of standardization among agencies in field methodology or data management. To be

  1. INCREMENTAL PRINCIPAL COMPONENT ANALYSIS BASED OUTLIER DETECTION METHODS FOR SPATIOTEMPORAL DATA STREAMS

    Directory of Open Access Journals (Sweden)

    A. Bhushan

    2015-07-01

    Full Text Available In this paper, we address outliers in spatiotemporal data streams obtained from sensors placed across geographically distributed locations. Outliers may appear in such sensor data due to various reasons such as instrumental error and environmental change. Real-time detection of these outliers is essential to prevent propagation of errors in subsequent analyses and results. Incremental Principal Component Analysis (IPCA is one possible approach for detecting outliers in such type of spatiotemporal data streams. IPCA has been widely used in many real-time applications such as credit card fraud detection, pattern recognition, and image analysis. However, the suitability of applying IPCA for outlier detection in spatiotemporal data streams is unknown and needs to be investigated. To fill this research gap, this paper contributes by presenting two new IPCA-based outlier detection methods and performing a comparative analysis with the existing IPCA-based outlier detection methods to assess their suitability for spatiotemporal sensor data streams.

  2. Open Source Initiative Powers Real-Time Data Streams

    Science.gov (United States)

    2014-01-01

    Under an SBIR contract with Dryden Flight Research Center, Creare Inc. developed a data collection tool called the Ring Buffered Network Bus. The technology has now been released under an open source license and is hosted by the Open Source DataTurbine Initiative. DataTurbine allows anyone to stream live data from sensors, labs, cameras, ocean buoys, cell phones, and more.

  3. Mining Frequent Item Sets in Asynchronous Transactional Data Streams over Time Sensitive Sliding Windows Model

    International Nuclear Information System (INIS)

    Javaid, Q.; Memon, F.; Talpur, S.; Arif, M.; Awan, M.D.

    2016-01-01

    EPs (Extracting Frequent Patterns) from the continuous transactional data streams is a challenging and critical task in some of the applications, such as web mining, data analysis and retail market, prediction and network monitoring, or analysis of stock market exchange data. Many algorithms have been developed previously for mining FPs (Frequent Patterns) from a data stream. Such algorithms are currently highly required to develop new solutions and approaches to the precise handling of data streams. New techniques, solutions, or approaches are developed to address unbounded, ordered, and continuous sequences of data and for the generation of data at a rapid speed from data streams. Hence, extracting FPs using fresh or recent data involves the high-level analysis of data streams. We have suggested an efficient technique for the window sliding model; this technique extracts new and fresh FPs from high-speed data streams. In this study, a CPILT (Compacted Tree Compact Pattern Tree) is developed to capture the latest contents in the stream and to efficiently remove outdated contents from the data stream. The main concept introduced in this work on CPILT is the dynamic restructuring of a tree, which is helpful in producing a compacted tree and the frequency descending structure of a tree on runtime. With the help of the mining technique of FP growth, a complete list of new and fresh FPs is obtained from a CPILT using an existing window. The memory usage and time complexity of the latest FPs in high-speed data streams can efficiently be determined through proper experimentation and analysis. (author)

  4. Indexing Density Models for Incremental Learning and Anytime Classification on Data Streams

    DEFF Research Database (Denmark)

    Seidl, Thomas; Assent, Ira; Kranen, Philipp

    2009-01-01

    Classification of streaming data faces three basic challenges: it has to deal with huge amounts of data, the varying time between two stream data items must be used best possible (anytime classification) and additional training data must be incrementally learned (anytime learning) for applying...... to the individual object to be classified) a hierarchy of mixture densities that represent kernel density estimators at successively coarser levels. Our probability density queries together with novel classification improvement strategies provide the necessary information for very effective classification at any...... point of interruption. Moreover, we propose a novel evaluation method for anytime classification using Poisson streams and demonstrate the anytime learning performance of the Bayes tree....

  5. SWAMP+: multiple subsequence alignment using associative massive parallelism

    Energy Technology Data Exchange (ETDEWEB)

    Steinfadt, Shannon Irene [Los Alamos National Laboratory; Baker, Johnnie W [KENT STATE UNIV.

    2010-10-18

    A new parallel algorithm SWAMP+ incorporates the Smith-Waterman sequence alignment on an associative parallel model known as ASC. It is a highly sensitive parallel approach that expands traditional pairwise sequence alignment. This is the first parallel algorithm to provide multiple non-overlapping, non-intersecting subsequence alignments with the accuracy of Smith-Waterman. The efficient algorithm provides multiple alignments similar to BLAST while creating a better workflow for the end users. The parallel portions of the code run in O(m+n) time using m processors. When m = n, the algorithmic analysis becomes O(n) with a coefficient of two, yielding a linear speedup. Implementation of the algorithm on the SIMD ClearSpeed CSX620 confirms this theoretical linear speedup with real timings.

  6. Hydrogeochemical and stream sediment reconnaissance basic data for Seguin NTMS quadrangle, Texas

    International Nuclear Information System (INIS)

    1978-01-01

    Results of a reconnaissance geochemical survey of the Sequin Quadrangle, Texas are reported. Field and laboratory data are presented for 848 groundwater, 950 stream sediment, and 406 stream water samples. Statistical and areal distributions of uranium and other possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided, and pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Groundwater data indicate that uranium concentrations above the 85th percentile occur along several northeast-southwest trends paralleling the regional strike of the major formations located within the survey area. The stream sediment data indicate that uranium is associated with heavy and/or resistate minerals in the Carrizo Sand and certain members of the Claiborne Group. Soluble uranium is primarily associated with the Cretaceous Formations, the Whitsett and Catahoula Formations, and sections of the Oakville and Fleming Formations. Stream water data corroborate well with both groundwater and stream sediment data. Anomalous values for uranium and associated pathfinder elements indicate that the Whitsett and Catahoula Formations and sections of the Oakville and Fleming Formations are potentially favorable for uranium mineralization. Anomalous values for certain pathfinder elements also occur in basins draining from the Beaumont Formation and may warrant further investigation

  7. Design for real-time data acquisition based on streaming technology

    International Nuclear Information System (INIS)

    Nakanishi, Hideya; Kojima, Mamoru

    2001-04-01

    For the LHD project a long-pulse plasma experiment of one-hour duration is planned. In this quasi steady-state operation, the data acquisition system will be required to continuously transfer the diagnostic data from the digitizer front-end and display them in real-time. The Compact PCI standard is used to replace the conventional CAMAC digitizers in LHD, because it provides good functionality for real-time data streaming and also a connectivity with modern PC technology. The digitizer scheme, interface to the host computer, adoption of data compression, and downstream applications are discussed in detail to design and implement this new real-time data streaming system for LHD plasma diagnostics. (author)

  8. Adaptive SVM for Data Stream Classification

    Directory of Open Access Journals (Sweden)

    Isah A. Lawal

    2017-07-01

    Full Text Available In this paper, we address the problem of learning an adaptive classifier for the classification of continuous streams of data. We present a solution based on incremental extensions of the Support Vector Machine (SVM learning paradigm that updates an existing SVM whenever new training data are acquired. To ensure that the SVM effectiveness is guaranteed while exploiting the newly gathered data, we introduce an on-line model selection approach in the incremental learning process. We evaluated the proposed method on real world applications including on-line spam email filtering and human action classification from videos. Experimental results show the effectiveness and the potential of the proposed approach.

  9. Stream on the Sky: Outsourcing Access Control Enforcement for Stream Data to the Cloud

    OpenAIRE

    Dinh, Tien Tuan Anh; Datta, Anwitaman

    2012-01-01

    There is an increasing trend for businesses to migrate their systems towards the cloud. Security concerns that arise when outsourcing data and computation to the cloud include data confidentiality and privacy. Given that a tremendous amount of data is being generated everyday from plethora of devices equipped with sensing capabilities, we focus on the problem of access controls over live streams of data based on triggers or sliding windows, which is a distinct and more challenging problem tha...

  10. Simulation and reconstruction of free-streaming data in CBM

    International Nuclear Information System (INIS)

    Friese, Volker

    2011-01-01

    The CBM experiment will investigate heavy-ion reactions at the FAIR facility at unprecedented interaction rates. This implies a novel read-out and data acquisition concept with self-triggered front-end electronics and free-streaming data. Event association must be performed in software on-line, and may require four-dimensional reconstruction routines. In order to study the problem of event association and to develop proper algorithms, simulations must be performed which go beyond the normal event-by-event processing as available from most experimental simulation frameworks. In this article, we discuss the challenges and concepts for the reconstruction of such free-streaming data and present first steps for a time-based simulation which is necessary for the development and validation of the reconstruction algorithms, and which requires modifications to the current software framework FAIRROOT as well as to the data model.

  11. Improved streaming analysis technique: spherical harmonics expansion of albedo data

    International Nuclear Information System (INIS)

    Albert, T.E.; Simmons, G.L.

    1979-01-01

    An improved albedo scattering technique was implemented with a three-dimensional Monte Carlo transport code for use in analyzing radiation streaming problems. The improvement was based on a shifted spherical Harmonics expansion of the doubly differential albedo data base. The result of the improvement was a factor of 3 to 10 reduction in data storage requirements and approximately a factor of 3 to 6 increase in computational speed. Comparisons of results obtained using the technique with measurements are shown for neutron streaming in one- and two-legged square concrete ducts

  12. Expanded stream gauging includes groundwater data and trends

    Science.gov (United States)

    Constantz, James E.; Barlow, Jeannie R.; Eddy-Miller, Cheryl; Caldwell, Rodney R.; Wheeler, Jerrod D.

    2012-01-01

    Population growth has increased water scarcity to the point that documenting current amounts of worldwide water resources is now as critical as any data collection in the Earth sciences. As a key element of this data collection, stream gauges yield continuous hydrologic information and document long-term trends, recording high-frequency hydrologic information over decadal to centennial time frames.

  13. StreamNet, annual report FY 2000; ANNUAL

    International Nuclear Information System (INIS)

    Schmidt, Bruce R.

    2001-01-01

    The StreamNet Project is a cooperative project that provides basic fishery management data in a consistent format across the Columbia Basin region, with some data from outside the region. Specific categories of data are acquired from the multiple data generating agencies in the Columbia Basin, converted into a standardized data exchange format (DEF) and distributed to fish researchers, managers and decision makers directly or through an on-line data retrieval system (www.streamnet.org). The project is funded by the Bonneville Power Administration (BPA) as part of the Northwest Power Planning Council's (NWPPC) Fish and Wildlife Program. This cooperative effort is composed of a region-wide project administered by the Pacific States Marine Fisheries Commission (PSMFC) that is responsible for project management, regional data management and data delivery (Region), plus seven contributing projects within the data generating entities: Columbia River Intertribal Fish Commission (CRITFC); Idaho Department of Fish and Game (IDFG); Montana Fish, Wildlife and Parks (MFWP); Oregon Department of Fish and Wildlife (ODFW); Shoshone-Bannock Tribes; U. S. Fish and Wildlife Service (FWS); and Washington Department of Fish and Wildlife (WDFW). The contributing projects are funded through the StreamNet contract but work within their respective agencies and are referred to here as the agency's StreamNet project (for example, ''IDFG StreamNet'' for Idaho's project). The StreamNet Project provides an important link in the chain of data flow in the Columbia Basin, with specific emphasis on data collected routinely over time by management agencies. Basic fish related data are collected in the field by the various state, tribal and federal agencies in the basin for purposes related to each agency's individual mission and responsibility. As a result, there often is a lack of standardization among agencies in field methodology or data management. To be able to utilize data for comparison or

  14. Real-time change detection in data streams with FPGAs

    International Nuclear Information System (INIS)

    Vega, J.; Dormido-Canto, S.; Cruz, T.; Ruiz, M.; Barrera, E.; Castro, R.; Murari, A.; Ochando, M.

    2014-01-01

    Highlights: • Automatic recognition of changes in data streams of multidimensional signals. • Detection algorithm based on testing exchangeability on-line. • Real-time and off-line applicability. • Real-time implementation in FPGAs. - Abstract: The automatic recognition of changes in data streams is useful in both real-time and off-line data analyses. This article shows several effective change-detecting algorithms (based on martingales) and describes their real-time applicability in the data acquisition systems through the use of Field Programmable Gate Arrays (FPGA). The automatic event recognition system is absolutely general and it does not depend on either the particular event to detect or the specific data representation (waveforms, images or multidimensional signals). The developed approach provides good results for change detection in both the temporal evolution of profiles and the two-dimensional spatial distribution of volume emission intensity. The average computation time in the FPGA is 210 μs per profile

  15. Recognition of periodic behavioral patterns from streaming mobility data

    NARCIS (Netherlands)

    Baratchi, Mitra; Meratnia, Nirvana; Havinga, Paul J.M.; Stojmenovic, Ivan; Cheng, Zixue; Guo, Song

    2014-01-01

    Ubiquitous location-aware sensing devices have facilitated collection of large volumes of mobility data streams from moving entities such as people and animals, among others. Extraction of various types of periodic behavioral patterns hidden in such large volume of mobility data helps in

  16. Streaming Pool: reuse, combine and create reactive streams with pleasure

    CERN Multimedia

    CERN. Geneva

    2017-01-01

    When connecting together heterogeneous and complex systems, it is not easy to exchange data between components. Streams of data are successfully used in industry in order to overcome this problem, especially in the case of "live" data. Streams are a specialization of the Observer design pattern and they provide asynchronous and non-blocking data flow. The ongoing effort of the ReactiveX initiative is one example that demonstrates how demanding this technology is even for big companies. Bridging the discrepancies of different technologies with common interfaces is already done by the Reactive Streams initiative and, in the JVM world, via reactive-streams-jvm interfaces. Streaming Pool is a framework for providing and discovering reactive streams. Through the mechanism of dependency injection provided by the Spring Framework, Streaming Pool provides a so called Discovery Service. This object can discover and chain streams of data that are technologically agnostic, through the use of Stream IDs. The stream to ...

  17. Method and apparatus of prefetching streams of varying prefetch depth

    Science.gov (United States)

    Gara, Alan [Mount Kisco, NY; Ohmacht, Martin [Yorktown Heights, NY; Salapura, Valentina [Chappaqua, NY; Sugavanam, Krishnan [Mahopac, NY; Hoenicke, Dirk [Seebruck-Seeon, DE

    2012-01-24

    Method and apparatus of prefetching streams of varying prefetch depth dynamically changes the depth of prefetching so that the number of multiple streams as well as the hit rate of a single stream are optimized. The method and apparatus in one aspect monitor a plurality of load requests from a processing unit for data in a prefetch buffer, determine an access pattern associated with the plurality of load requests and adjust a prefetch depth according to the access pattern.

  18. A fast calculating two-stream-like multiple scattering algorithm that captures azimuthal and elevation variations

    Science.gov (United States)

    Fiorino, Steven T.; Elmore, Brannon; Schmidt, Jaclyn; Matchefts, Elizabeth; Burley, Jarred L.

    2016-05-01

    Properly accounting for multiple scattering effects can have important implications for remote sensing and possibly directed energy applications. For example, increasing path radiance can affect signal noise. This study describes the implementation of a fast-calculating two-stream-like multiple scattering algorithm that captures azimuthal and elevation variations into the Laser Environmental Effects Definition and Reference (LEEDR) atmospheric characterization and radiative transfer code. The multiple scattering algorithm fully solves for molecular, aerosol, cloud, and precipitation single-scatter layer effects with a Mie algorithm at every calculation point/layer rather than an interpolated value from a pre-calculated look-up-table. This top-down cumulative diffusivity method first considers the incident solar radiance contribution to a given layer accounting for solid angle and elevation, and it then measures the contribution of diffused energy from previous layers based on the transmission of the current level to produce a cumulative radiance that is reflected from a surface and measured at the aperture at the observer. Then a unique set of asymmetry and backscattering phase function parameter calculations are made which account for the radiance loss due to the molecular and aerosol constituent reflectivity within a level and allows for a more accurate characterization of diffuse layers that contribute to multiple scattered radiances in inhomogeneous atmospheres. The code logic is valid for spectral bands between 200 nm and radio wavelengths, and the accuracy is demonstrated by comparing the results from LEEDR to observed sky radiance data.

  19. Mining top-k frequent closed itemsets in data streams using sliding window

    International Nuclear Information System (INIS)

    Rehman, Z.; Shahbaz, M.

    2013-01-01

    Frequent itemset mining has become a popular research area in data mining community since the last few years. T here are two main technical hitches while finding frequent itemsets. First, to provide an appropriate minimum support value to start and user need to tune this minimum support value by running the algorithm again and again. Secondly, generated frequent itemsets are mostly numerous and as a result a number of association rules generated are also very large in numbers. Applications dealing with streaming environment need to process the data received at high rate, therefore, finding frequent itemsets in data streams becomes complex. In this paper, we present an algorithm to mine top-k frequent closed itemsets using sliding window approach from streaming data. We developed a single-pass algorithm to find frequent closed itemsets of length between user's defined minimum and maximum- length. To improve the performance of algorithm and to avoid rescanning of data, we have transformed data into bitmap based tree data structure. (author)

  20. Scalable IC Platform for Smart Cameras

    Directory of Open Access Journals (Sweden)

    Harry Broers

    2005-08-01

    Full Text Available Smart cameras are among the emerging new fields of electronics. The points of interest are in the application areas, software and IC development. In order to reduce cost, it is worthwhile to invest in a single architecture that can be scaled for the various application areas in performance (and resulting power consumption. In this paper, we show that the combination of an SIMD (single-instruction multiple-data processor and a general-purpose DSP is very advantageous for the image processing tasks encountered in smart cameras. While the SIMD processor gives the very high performance necessary by exploiting the inherent data parallelism found in the pixel crunching part of the algorithms, the DSP offers a friendly approach to the more complex tasks. The paper continues to motivate that SIMD processors have very convenient scaling properties in silicon, making the complete, SIMD-DSP architecture suitable for different application areas without changing the software suite. Analysis of the changes in power consumption due to scaling shows that for typical image processing tasks, it is beneficial to scale the SIMD processor to use the maximum level of parallelism available in the algorithm if the IC supply voltage can be lowered. If silicon cost is of importance, the parallelism of the processor should be scaled to just reach the desired performance given the speed of the silicon.

  1. CC_TRS: Continuous Clustering of Trajectory Stream Data Based on Micro Cluster Life

    Directory of Open Access Journals (Sweden)

    Musaab Riyadh

    2017-01-01

    Full Text Available The rapid spreading of positioning devices leads to the generation of massive spatiotemporal trajectories data. In some scenarios, spatiotemporal data are received in stream manner. Clustering of stream data is beneficial for different applications such as traffic management and weather forecasting. In this article, an algorithm for Continuous Clustering of Trajectory Stream Data Based on Micro Cluster Life is proposed. The algorithm consists of two phases. There is the online phase where temporal micro clusters are used to store summarized spatiotemporal information for each group of similar segments. The clustering task in online phase is based on temporal micro cluster lifetime instead of time window technique which divides stream data into time bins and clusters each bin separately. For offline phase, a density based clustering approach is used to generate macro clusters depending on temporal micro clusters. The evaluation of the proposed algorithm on real data sets shows the efficiency and the effectiveness of the proposed algorithm and proved it is efficient alternative to time window technique.

  2. Longitudinal structure in temperate stream fish communities: evaluating conceptual models with temporal data

    Science.gov (United States)

    Roberts, James H.; Hitt, Nathaniel P.

    2010-01-01

    Five conceptual models of longitudinal fish community organization in streams were examined: (1) niche diversity model (NDM), (2) stream continuum model (SCM), (3) immigrant accessibility model (IAM), (4) environmental stability model (ESM), and (5) adventitious stream model (ASM). We used differences among models in their predictions about temporal species turnover, along with five spatiotemporal fish community data sets, to evaluate model applicability. Models were similar in predicting a positive species richness–stream size relationship and longitudinal species nestedness, but differed in predicting either similar temporal species turnover throughout the stream continuum (NDM, SCM), higher turnover upstream (IAM, ESM), or higher turnover downstream (ASM). We calculated measures of spatial and temporal variation from spatiotemporal fish data in five wadeable streams in central and eastern North America spanning 34–68 years (French Creek [New York], Piasa Creek [Illinois], Spruce Run [Virginia], Little Stony Creek [Virginia], and Sinking Creek [Virginia]). All streams exhibited substantial species turnover (i.e., at least 27% turnover in stream-scale species pools), in contrast to the predictions of the SCM. Furthermore, community change was greater in downstream than upstream reaches in four of five streams. This result is most consistent with the ASM and suggests that downstream communities are strongly influenced by migrants to and from species pools outside the focal stream. In Sinking Creek, which is isolated from external species pools, temporal species turnover (via increased richness) was higher upstream than downstream, which is a pattern most consistent with the IAM or ESM. These results corroborate the hypothesis that temperate stream habitats and fish communities are temporally dynamic and that fish migration and environmental disturbances play fundamental roles in stream fish community organization.

  3. Mining Twitter Data Stream to Augment NASA GPM Validation

    Science.gov (United States)

    Teng, W. L.; Albayrak, A.; Huffman, G. J.; Vollmer, B.

    2017-12-01

    The Twitter data stream is an important new source of real-time and historical global information for potentially augmenting the validation program of NASA's Global Precipitation Measurement (GPM) mission. There have been other similar uses of Twitter, though mostly related to natural hazards monitoring and management. The validation of satellite precipitation estimates is challenging, because many regions lack data or access to data, especially outside of the U.S. and in remote and developing areas. The time-varying set of "precipitation" tweets can be thought of as an organic network of rain gauges, potentially providing a widespread view of precipitation occurrence. Twitter provides a large source of crowd for crowdsourcing. During a 24-hour period in the middle of the snow storm this past March in the U.S. Northeast, we collected more than 13,000 relevant precipitation tweets with exact geolocation. The overall objective of our project is to determine the extent to which processed tweets can provide additional information that improves the validation of GPM data. Though our current effort focuses on tweets and precipitation, our approach is general and applicable to other social media and other geophysical measurements. Specifically, we have developed an operational infrastructure for processing tweets, in a format suitable for analysis with GPM data; engaged with potential participants, both passive and active, to "enrich" the Twitter stream; and inter-compared "precipitation" tweet data, ground station data, and GPM retrievals. In this presentation, we detail the technical capabilities of our tweet processing infrastructure, including data abstraction, feature extraction, search engine, context-awareness, real-time processing, and high volume (big) data processing; various means for "enriching" the Twitter stream; and results of inter-comparisons. Our project should bring a new kind of visibility to Twitter and engender a new kind of appreciation of the value

  4. Observed TRU data from nuclear utility waste streams

    International Nuclear Information System (INIS)

    Wessman, R.A.; Floyd, J.G.; Leventhal, L.

    1990-01-01

    TMA/Norcal has performed 10CFR61 analysis of radioactive waste streams from BWR's and PWR's since 1983. Many standard and non-routine sample types have been received for analysis from nuclear power plants nation-wide. In addition to the 10CFR61 Tables I and II analyses, we also have analyzed for many of the supplementary isotopes. As part of this program TRU analyses are required. As a result, have accumulated a significant amount of data for plutonium, americium, and curium in radioactive waste for many different sample matrices from many different waste streams. This paper will present our analytical program for 10CFR61 TRU. The laboratory methodology including chemical and radiometric procedures is discussed. The sensitivity of our measurements and ability to meet the lower limits of detection is also discussed. Secondly, a review of TRU data is presented. Scaling factors and their ranges from selected PWR stations are included. We discuss some features of, and limits to, interpretation of these data. 8 refs., 3 tabs

  5. Assessing the influence of multiple stressors on stream diatom metrics in the upper Midwest, USA

    Science.gov (United States)

    Munn, Mark D.; Waite, Ian R.; Konrad, Christopher P.

    2018-01-01

    Water resource managers face increasing challenges in identifying what physical and chemical stressors are responsible for the alteration of biological conditions in streams. The objective of this study was to assess the comparative influence of multiple stressors on benthic diatoms at 98 sites that spanned a range of stressors in an agriculturally dominated region in the upper Midwest, USA. The primary stressors of interest included: nutrients, herbicides and fungicides, sediment, and streamflow; although the influence of physical habitat was incorporated in the assessment. Boosted Regression Tree was used to examine both the sensitivity of various diatom metrics and the relative importance of the primary stressors. Percent Sensitive Taxa, percent Highly Motile Taxa, and percent High Phosphorus Taxa had the strongest response to stressors. Habitat and total phosphorous were the most common discriminators of diatom metrics, with herbicides as secondary factors. A Classification and Regression Tree (CART) model was used to examine conditional relations among stressors and indicated that fine-grain streams had a lower percentage of Sensitive Taxa than coarse-grain streams, with Sensitive Taxa decreasing further with increased water temperature (>30 °C) and triazine concentrations (>1500 ng/L). In contrast, streams dominated by coarse-grain substrate contained a higher percentage of Sensitive Taxa, with relative abundance increasing with lower water temperatures (water depth (water temperature appears to be a major limiting factor in Midwest streams; whereas both total phosphorus and percent fines showed a slight subsidy-stress response. While using benthic algae for assessing stream quality can be challenging, field-based studies can elucidate stressor effects and interactions when the response variables are appropriate, sufficient stressor resolution is achieved, and the number and type of sites represent a gradient of stressor conditions and at least a quasi

  6. PROXY-BASED PATCHING STREAM TRANSMISSION STRATEGY IN MOBILE STREAMING MEDIA SYSTEM

    Institute of Scientific and Technical Information of China (English)

    Liao Jianxin; Lei Zhengxiong; Ma Xutao; Zhu Xiaomin

    2006-01-01

    A mobile transmission strategy, PMPatching (Proxy-based Mobile Patching) transmission strategy is proposed, it applies to the proxy-based mobile streaming media system in Wideband Code Division Multiple Access (WCDMA) network. Performance of the whole system can be improved by using patching stream to transmit anterior part of the suffix that had been played back, and by batching all the demands for the suffix arrived in prefix period and patching stream transmission threshold period. Experimental results show that this strategy can efficiently reduce average network transmission cost and number of channels consumed in central streaming media server.

  7. Guidelines and Procedures for Computing Time-Series Suspended-Sediment Concentrations and Loads from In-Stream Turbidity-Sensor and Streamflow Data

    Science.gov (United States)

    Rasmussen, Patrick P.; Gray, John R.; Glysson, G. Douglas; Ziegler, Andrew C.

    2009-01-01

    In-stream continuous turbidity and streamflow data, calibrated with measured suspended-sediment concentration data, can be used to compute a time series of suspended-sediment concentration and load at a stream site. Development of a simple linear (ordinary least squares) regression model for computing suspended-sediment concentrations from instantaneous turbidity data is the first step in the computation process. If the model standard percentage error (MSPE) of the simple linear regression model meets a minimum criterion, this model should be used to compute a time series of suspended-sediment concentrations. Otherwise, a multiple linear regression model using paired instantaneous turbidity and streamflow data is developed and compared to the simple regression model. If the inclusion of the streamflow variable proves to be statistically significant and the uncertainty associated with the multiple regression model results in an improvement over that for the simple linear model, the turbidity-streamflow multiple linear regression model should be used to compute a suspended-sediment concentration time series. The computed concentration time series is subsequently used with its paired streamflow time series to compute suspended-sediment loads by standard U.S. Geological Survey techniques. Once an acceptable regression model is developed, it can be used to compute suspended-sediment concentration beyond the period of record used in model development with proper ongoing collection and analysis of calibration samples. Regression models to compute suspended-sediment concentrations are generally site specific and should never be considered static, but they represent a set period in a continually dynamic system in which additional data will help verify any change in sediment load, type, and source.

  8. Process recognition in multi-element soil and stream-sediment geochemical data

    Science.gov (United States)

    Grunsky, E.C.; Drew, L.J.; Sutphin, D.M.

    2009-01-01

    Stream-sediment and soil geochemical data from the Upper and Lower Coastal Plains of South Carolina (USA) were studied to determine relationships between soils and stream sediments. From multi-element associations, characteristic compositions were determined for both media. Primary associations of elements reflect mineralogy, including heavy minerals, carbonates and clays, and the effects of groundwater. The effects of groundwater on element concentrations are more evident in soils than stream sediments. A "winnowing index" was created using ratios of Th to Al that revealed differing erosional and depositional environments. Both soils and stream sediments from the Upper and Lower Coastal Plains show derivation from similar materials and subsequent similar multi-element relationships, but have some distinct differences. In the Lower Coastal Plain, soils have high values of elements concentrated in heavy minerals (Ce, Y, Th) that grade into high values of elements concentrated into finer-grain-size, lower-density materials, primarily comprised of carbonates and feldspar minerals (Mg, Ca, Na, K, Al). These gradational trends in mineralogy and geochemistry are inferred to reflect reworking of materials during marine transgressions and regressions. Upper Coastal Plain stream-sediment geochemistry shows a higher winnowing index relative to soil geochemistry. A comparison of the 4 media (Upper Coastal Plain soils and stream sediments and Lower Coastal Plain soils and stream sediments) shows that Upper Coastal Plain stream sediments have a higher winnowing index and a higher concentration of elements contained within heavy minerals, whereas Lower Coastal Plain stream sediments show a strong correlation between elements typically contained within clays. It is not possible to calculate a functional relationship between stream sediment-soil compositions for all elements due to the complex history of weathering, deposition, reworking and re-deposition. However, depending on

  9. Efficient Processing of Continuous Skyline Query over Smarter Traffic Data Stream for Cloud Computing

    Directory of Open Access Journals (Sweden)

    Wang Hanning

    2013-01-01

    Full Text Available The analyzing and processing of multisource real-time transportation data stream lay a foundation for the smart transportation's sensibility, interconnection, integration, and real-time decision making. Strong computing ability and valid mass data management mode provided by the cloud computing, is feasible for handling Skyline continuous query in the mass distributed uncertain transportation data stream. In this paper, we gave architecture of layered smart transportation about data processing, and we formalized the description about continuous query over smart transportation data Skyline. Besides, we proposed mMR-SUDS algorithm (Skyline query algorithm of uncertain transportation stream data based on micro-batchinMap Reduce based on sliding window division and architecture.

  10. SmartCell: An Energy Efficient Coarse-Grained Reconfigurable Architecture for Stream-Based Applications

    Directory of Open Access Journals (Sweden)

    Liang Cao

    2009-01-01

    Full Text Available This paper presents SmartCell, a novel coarse-grained reconfigurable architecture, which tiles a large number of processor elements with reconfigurable interconnection fabrics on a single chip. SmartCell is able to provide high performance and energy efficient processing for stream-based applications. It can be configured to operate in various modes, such as SIMD, MIMD, and systolic array. This paper describes the SmartCell architecture design, including processing element, reconfigurable interconnection fabrics, instruction and control process, and configuration scheme. The SmartCell prototype with 64 PEs is implemented using 0.13  m CMOS standard cell technology. The core area is about 8.5  , and the power consumption is about 1.6 mW/MHz. The performance is evaluated through a set of benchmark applications, and then compared with FPGA, ASIC, and two well-known reconfigurable architectures including RaPiD and Montium. The results show that the SmartCell can bridge the performance and flexibility gap between ASIC and FPGA. It is also about 8% and 69% more energy efficient than Montium and RaPiD systems for evaluated benchmarks. Meanwhile, SmartCell can achieve 4 and 2 times more throughput gains when comparing with Montium and RaPiD, respectively. It is concluded that SmartCell system is a promising reconfigurable and energy efficient architecture for stream processing.

  11. DataCell: Exploiting the Power of Relational Databases for Efficient Stream Processing

    NARCIS (Netherlands)

    E. Liarou (Erietta); M.L. Kersten (Martin)

    2009-01-01

    htmlabstractDesigned for complex event processing, DataCell is a research prototype database system in the area of sensor stream systems. Under development at CWI, it belongs to the MonetDB database system family. CWI researchers innovatively built a stream engine directly on top of a database

  12. Bivariate functional data clustering: grouping streams based on a varying coefficient model of the stream water and air temperature relationship

    Science.gov (United States)

    H. Li; X. Deng; Andy Dolloff; E. P. Smith

    2015-01-01

    A novel clustering method for bivariate functional data is proposed to group streams based on their water–air temperature relationship. A distance measure is developed for bivariate curves by using a time-varying coefficient model and a weighting scheme. This distance is also adjusted by spatial correlation of streams via the variogram. Therefore, the proposed...

  13. Efficient Streaming Mass Spatio-Temporal Vehicle Data Access in Urban Sensor Networks Based on Apache Storm.

    Science.gov (United States)

    Zhou, Lianjie; Chen, Nengcheng; Chen, Zeqiang

    2017-04-10

    The efficient data access of streaming vehicle data is the foundation of analyzing, using and mining vehicle data in smart cities, which is an approach to understand traffic environments. However, the number of vehicles in urban cities has grown rapidly, reaching hundreds of thousands in number. Accessing the mass streaming data of vehicles is hard and takes a long time due to limited computation capability and backward modes. We propose an efficient streaming spatio-temporal data access based on Apache Storm (ESDAS) to achieve real-time streaming data access and data cleaning. As a popular streaming data processing tool, Apache Storm can be applied to streaming mass data access and real time data cleaning. By designing the Spout/bolt workflow of topology in ESDAS and by developing the speeding bolt and other bolts, Apache Storm can achieve the prospective aim. In our experiments, Taiyuan BeiDou bus location data is selected as the mass spatio-temporal data source. In the experiments, the data access results with different bolts are shown in map form, and the filtered buses' aggregation forms are different. In terms of performance evaluation, the consumption time in ESDAS for ten thousand records per second for a speeding bolt is approximately 300 milliseconds, and that for MongoDB is approximately 1300 milliseconds. The efficiency of ESDAS is approximately three times higher than that of MongoDB.

  14. Efficient Streaming Mass Spatio-Temporal Vehicle Data Access in Urban Sensor Networks Based on Apache Storm

    Directory of Open Access Journals (Sweden)

    Lianjie Zhou

    2017-04-01

    Full Text Available The efficient data access of streaming vehicle data is the foundation of analyzing, using and mining vehicle data in smart cities, which is an approach to understand traffic environments. However, the number of vehicles in urban cities has grown rapidly, reaching hundreds of thousands in number. Accessing the mass streaming data of vehicles is hard and takes a long time due to limited computation capability and backward modes. We propose an efficient streaming spatio-temporal data access based on Apache Storm (ESDAS to achieve real-time streaming data access and data cleaning. As a popular streaming data processing tool, Apache Storm can be applied to streaming mass data access and real time data cleaning. By designing the Spout/bolt workflow of topology in ESDAS and by developing the speeding bolt and other bolts, Apache Storm can achieve the prospective aim. In our experiments, Taiyuan BeiDou bus location data is selected as the mass spatio-temporal data source. In the experiments, the data access results with different bolts are shown in map form, and the filtered buses’ aggregation forms are different. In terms of performance evaluation, the consumption time in ESDAS for ten thousand records per second for a speeding bolt is approximately 300 milliseconds, and that for MongoDB is approximately 1300 milliseconds. The efficiency of ESDAS is approximately three times higher than that of MongoDB.

  15. Capabilities of Raspberry Pi 2 for Big Data and Video Streaming Applications in Data Centres

    NARCIS (Netherlands)

    Schot, Nick J.; Velthuis, Paul J.E.; Postema, Björn Frits; Remke, Anne Katharina Ingrid; Remke, A.K.I.; Haverkort, Boudewijn R.H.M.; Haverkort, B.R.H.M.

    Many new data centres have been built in recent years in order to keep up with the rising demand for server capacity. These data centres require a lot of electrical energy and cooling. Big data and video streaming are two heavily used applications in data centres. This paper experimentally

  16. Multiple bio-monitoring system using visible light for electromagnetic-wave free indoor healthcare

    Science.gov (United States)

    An, Jinyoung; Pham, Ngoc Quan; Chung, Wan-Young

    2017-12-01

    In this paper, a multiple biomedical data transmission system with visible light communication (VLC) is proposed for an electromagnetic-wave-free indoor healthcare. VLC technology has emerged as an alternative solution to radio-frequency (RF) wireless systems, due to its various merits, e.g., ubiquity, power efficiency, no RF radiation, and security. With VLC, critical bio-medical signals, including electrocardiography (ECG), can be transmitted in places where RF radiation is restricted. This potential advantage of VLC could save more lives in emergency situations. A time hopping (TH) scheme is employed to transfer multiple medical-data streams in real time with a simple system design. Multiple data streams are transmitted using identical color LEDs and go into an optical detector. The received multiple data streams are demodulated and rearranged using a TH-based demodulator. The medical data is then monitored and managed to provide the necessary medical care for each patient.

  17. Cooperative Coding and Caching for Streaming Data in Multihop Wireless Networks

    Directory of Open Access Journals (Sweden)

    Liu Jiangchuan

    2010-01-01

    Full Text Available This paper studies the distributed caching managements for the current flourish of the streaming applications in multihop wireless networks. Many caching managements to date use randomized network coding approach, which provides an elegant solution for ubiquitous data accesses in such systems. However, the encoding, essentially a combination operation, makes the coded data difficult to be changed. In particular, to accommodate new data, the system may have to first decode all the combined data segments, remove some unimportant ones, and then reencode the data segments again. This procedure is clearly expensive for continuously evolving data storage. As such, we introduce a novel Cooperative Coding and Caching ( scheme, which allows decoding-free data removal through a triangle-like codeword organization. Its decoding performance is very close to the conventional network coding with only a sublinear overhead. Our scheme offers a promising solution to the caching management for streaming data.

  18. A high performance data parallel tensor contraction framework: Application to coupled electro-mechanics

    Science.gov (United States)

    Poya, Roman; Gil, Antonio J.; Ortigosa, Rogelio

    2017-07-01

    The paper presents aspects of implementation of a new high performance tensor contraction framework for the numerical analysis of coupled and multi-physics problems on streaming architectures. In addition to explicit SIMD instructions and smart expression templates, the framework introduces domain specific constructs for the tensor cross product and its associated algebra recently rediscovered by Bonet et al. (2015, 2016) in the context of solid mechanics. The two key ingredients of the presented expression template engine are as follows. First, the capability to mathematically transform complex chains of operations to simpler equivalent expressions, while potentially avoiding routes with higher levels of computational complexity and, second, to perform a compile time depth-first or breadth-first search to find the optimal contraction indices of a large tensor network in order to minimise the number of floating point operations. For optimisations of tensor contraction such as loop transformation, loop fusion and data locality optimisations, the framework relies heavily on compile time technologies rather than source-to-source translation or JIT techniques. Every aspect of the framework is examined through relevant performance benchmarks, including the impact of data parallelism on the performance of isomorphic and nonisomorphic tensor products, the FLOP and memory I/O optimality in the evaluation of tensor networks, the compilation cost and memory footprint of the framework and the performance of tensor cross product kernels. The framework is then applied to finite element analysis of coupled electro-mechanical problems to assess the speed-ups achieved in kernel-based numerical integration of complex electroelastic energy functionals. In this context, domain-aware expression templates combined with SIMD instructions are shown to provide a significant speed-up over the classical low-level style programming techniques.

  19. Geospatial Image Stream Processing: Models, techniques, and applications in remote sensing change detection

    Science.gov (United States)

    Rueda-Velasquez, Carlos Alberto

    Detection of changes in environmental phenomena using remotely sensed data is a major requirement in the Earth sciences, especially in natural disaster related scenarios where real-time detection plays a crucial role in the saving of human lives and the preservation of natural resources. Although various approaches formulated to model multidimensional data can in principle be applied to the inherent complexity of remotely sensed geospatial data, there are still challenging peculiarities that demand a precise characterization in the context of change detection, particularly in scenarios of fast changes. In the same vein, geospatial image streams do not fit appropriately in the standard Data Stream Management System (DSMS) approach because these systems mainly deal with tuple-based streams. Recognizing the necessity for a systematic effort to address the above issues, the work presented in this thesis is a concrete step toward the foundation and construction of an integrated Geospatial Image Stream Processing framework, GISP. First, we present a data and metadata model for remotely sensed image streams. We introduce a precise characterization of images and image streams in the context of remotely sensed geospatial data. On this foundation, we define spatially-aware temporal operators with a consistent semantics for change analysis tasks. We address the change detection problem in settings where multiple image stream sources are available, and thus we introduce an architectural design for the processing of geospatial image streams from multiple sources. With the aim of targeting collaborative scientific environments, we construct a realization of our architecture based on Kepler, a robust and widely used scientific workflow management system, as the underlying computational support; and open data and Web interface standards, as a means to facilitate the interoperability of GISP instances with other processing infrastructures and client applications. We demonstrate our

  20. Crowdsourcing Stream Stage in Data Scarce Regions: Applications of CrowdHydrology

    Science.gov (United States)

    Lowry, C.; Fienen, M. N.

    2013-12-01

    Crowdsourced data collection using citizen scientists and mobile phones is a promising way to collect supplemental information in data scarce or remote regions. The research presented here explore the possibilities and pitfalls of crowdsourcing hydrologic data via mobile phone text messaging through the example of CrowdHydrology, a distributed network of over 40 stream gages in four states. Signage at the CrowdHydrology gages ask citizen scientists to answer to a simple question via text message: 'What is the water height?'. While these data in no way replace more traditional measurements of stream stage, they do provide low cost supplemental measurements in data scarce regions. Results demonstrate the accuracy of crowdsourced data and provide insight for successful future crowdsourced data collection efforts. A less recognized benefit is that even in data rich areas, crowdsourced data collection is a cost-effective way to perform quality assurance on more sophisticated, and costly, data collection efforts.

  1. Data quality objectives for the B-Cell waste stream classification sampling

    International Nuclear Information System (INIS)

    Barnett, J.M.

    1998-01-01

    This document defines the data quality objectives, (DQOS) for sampling the B-Cell racks waste stream. The sampling effort is concentrated on determining a ratio of Cs-137 to Sr-90 and Cs-137 to transuranics (TRU). Figure 1.0 shows the logic path of sampling effort. The flow chart begins with sample and data acquisition and progresses toward (a) statistical confidence and waste classification boundaries, (b) management decisions based on the input parameters and technical methods available, and (c) grout container volume/weight limits and radiation limits. The end result will be accurately classifying the B-Cell rack waste stream

  2. Enhancing Network Data Obliviousness in Trusted Execution Environment-based Stream Processing Systems

    KAUST Repository

    Alsibyani, Hassan M.

    2018-05-15

    Cloud computing usage is increasing and a common concern is the privacy and security of the data and computation. Third party cloud environments are not considered fit for processing private information because the data will be revealed to the cloud provider. However, Trusted Execution Environments (TEEs), such as Intel SGX, provide a way for applications to run privately and securely on untrusted platforms. Nonetheless, using a TEE by itself for stream processing systems is not sufficient since network communication patterns may leak properties of the data under processing. This work addresses leaky topology structures and suggests mitigation techniques for each of these. We create specific metrics to evaluate leaks occurring from the network patterns; the metrics measure information leaked when the stream processing system is running. We consider routing techniques for inter-stage communication in a streaming application to mitigate this data leakage. We consider a dynamic policy to change the mitigation technique depending on how much information is currently leaking. Additionally, we consider techniques to hide irregularities resulting from a filtering stage in a topology. We also consider leakages resulting from applications containing cycles. For each of the techniques, we explore their effectiveness in terms of the advantage they provide in overcoming the network leakage. The techniques are tested partly using simulations and some were implemented in a prototype SGX-based stream processing system.

  3. Implicit Unstructured Aerodynamics on Emerging Multi- and Many-Core HPC Architectures

    KAUST Repository

    Al Farhan, Mohammed A.; Kaushik, Dinesh K.; Keyes, David E.

    2017-01-01

    Instruction, Multiple Data (SIMD) for hundreds of threads per node. We explore thread-level performance optimizations on state-of-the-art multi- and many-core Intel processors, including the second generation of Xeon Phi, Knights Landing (KNL). We study

  4. Quantifying in-stream nitrate reaction rates using continuously-collected water quality data

    Science.gov (United States)

    Matthew Miller; Anthony Tesoriero; Paul Capel

    2016-01-01

    High frequency in situ nitrate data from three streams of varying hydrologic condition, land use, and watershed size were used to quantify the mass loading of nitrate to streams from two sources – groundwater discharge and event flow – at a daily time step for one year. These estimated loadings were used to quantify temporally-variable in-stream nitrate processing ...

  5. Cooperative Coding and Caching for Streaming Data in Multihop Wireless Networks

    Directory of Open Access Journals (Sweden)

    Dan Wang

    2010-01-01

    Full Text Available This paper studies the distributed caching managements for the current flourish of the streaming applications in multihop wireless networks. Many caching managements to date use randomized network coding approach, which provides an elegant solution for ubiquitous data accesses in such systems. However, the encoding, essentially a combination operation, makes the coded data difficult to be changed. In particular, to accommodate new data, the system may have to first decode all the combined data segments, remove some unimportant ones, and then reencode the data segments again. This procedure is clearly expensive for continuously evolving data storage. As such, we introduce a novel Cooperative Coding and Caching (C3 scheme, which allows decoding-free data removal through a triangle-like codeword organization. Its decoding performance is very close to the conventional network coding with only a sublinear overhead. Our scheme offers a promising solution to the caching management for streaming data.

  6. FORTRAN computer programs to process Savannah River Laboratory hydrogeochemical and stream-sediment reconnaissance data

    International Nuclear Information System (INIS)

    Zinkl, R.J.; Shettel, D.L. Jr.; D'Andrea, R.F. Jr.

    1980-03-01

    FORTRAN computer programs have been written to read, edit, and reformat the hydrogeochemical and stream-sediment reconnaissance data produced by Savannah River Laboratory for the National Uranium Resource Evaluation program. The data are presorted by Savannah River Laboratory into stream sediment, ground water, and stream water for each 1 0 x 2 0 quadrangle. Extraneous information is eliminated, and missing analyses are assigned a specific value (-99999.0). Negative analyses are below the detection limit; the absolute value of a negative analysis is assumed to be the detection limit

  7. Approximating Frequent Items in Asynchronous Data Stream over a Sliding Window

    Directory of Open Access Journals (Sweden)

    Ho-Leung Chan

    2011-09-01

    Full Text Available In an asynchronous data stream, the data items may be out of order with respect to their original timestamps. This paper studies the space complexity required by a data structure to maintain such a data stream so that it can approximate the set of frequent items over a sliding time window with sufficient accuracy. Prior to our work, the best solution is given by Cormode et al. [1], who gave an O (1/ε log W log (εB/ log W min {log W, 1/ε} log |U|- space data structure that can approximate the frequent items within an ε error bound, where W and B are parameters of the sliding window, and U is the set of all possible item names. We gave a more space-efficient data structure that only requires O (1/ε log W log (εB/ logW log log W space.

  8. Self-adaptive change detection in streaming data with non-stationary distribution

    KAUST Repository

    Zhang, Xiangliang; Wang, Wei

    2010-01-01

    Non-stationary distribution, in which the data distribution evolves over time, is a common issue in many application fields, e.g., intrusion detection and grid computing. Detecting the changes in massive streaming data with a non

  9. A multimetric approach for predicting the ecological integrity of New Zealand streams

    Directory of Open Access Journals (Sweden)

    Clapcott J.E.

    2014-01-01

    Full Text Available Integrating multiple measures of stream health into a combined metric can provide a holistic assessment of the ecological integrity of a stream. The aim of this study was to develop a multimetric index (MMI of stream integrity based on predictive modelling of national data sets of water quality, macroinvertebrates, fish and ecosystem process metrics. We used a boosted regression tree approach to calculate an observed/expected score for each metric prior to combining metrics in a MMI based on data availability and the strength of predictive models. The resulting MMI provides a geographically meaningful prediction of the ecological integrity of rivers in New Zealand, but identifies limitations in data and approach, providing focus for ongoing research.

  10. StreamMap: Smooth Dynamic Visualization of High-Density Streaming Points.

    Science.gov (United States)

    Li, Chenhui; Baciu, George; Han, Yu

    2018-03-01

    Interactive visualization of streaming points for real-time scatterplots and linear blending of correlation patterns is increasingly becoming the dominant mode of visual analytics for both big data and streaming data from active sensors and broadcasting media. To better visualize and interact with inter-stream patterns, it is generally necessary to smooth out gaps or distortions in the streaming data. Previous approaches either animate the points directly or present a sampled static heat-map. We propose a new approach, called StreamMap, to smoothly blend high-density streaming points and create a visual flow that emphasizes the density pattern distributions. In essence, we present three new contributions for the visualization of high-density streaming points. The first contribution is a density-based method called super kernel density estimation that aggregates streaming points using an adaptive kernel to solve the overlapping problem. The second contribution is a robust density morphing algorithm that generates several smooth intermediate frames for a given pair of frames. The third contribution is a trend representation design that can help convey the flow directions of the streaming points. The experimental results on three datasets demonstrate the effectiveness of StreamMap when dynamic visualization and visual analysis of trend patterns on streaming points are required.

  11. Real-Time Joint Streaming Data Processing from Social and Physical Sensors

    Science.gov (United States)

    Kropivnitskaya, Y. Y.; Qin, J.; Tiampo, K. F.; Bauer, M.

    2014-12-01

    The results of the technological breakthroughs in computing that have taken place over the last few decades makes it possible to achieve emergency management objectives that focus on saving human lives and decreasing economic effects. In particular, the integration of a wide variety of information sources, including observations from spatially-referenced physical sensors and new social media sources, enables better real-time seismic hazard analysis through distributed computing networks. The main goal of this work is to utilize innovative computational algorithms for better real-time seismic risk analysis by integrating different data sources and processing tools into streaming and cloud computing applications. The Geological Survey of Canada operates the Canadian National Seismograph Network (CNSN) with over 100 high-gain instruments and 60 low-gain or strong motion seismographs. The processing of the continuous data streams from each station of the CNSN provides the opportunity to detect possible earthquakes in near real-time. The information from physical sources is combined to calculate a location and magnitude for an earthquake. The automatically calculated results are not always sufficiently precise and prompt that can significantly reduce the response time to a felt or damaging earthquake. Social sensors, here represented as Twitter users, can provide information earlier to the general public and more rapidly to the emergency planning and disaster relief agencies. We introduce joint streaming data processing from social and physical sensors in real-time based on the idea that social media observations serve as proxies for physical sensors. By using the streams of data in the form of Twitter messages, each of which has an associated time and location, we can extract information related to a target event and perform enhanced analysis by combining it with physical sensor data. Results of this work suggest that the use of data from social media, in conjunction

  12. Studying the effect of the shape parameters on the performance of the darrieus wind turbine using the multiple double disk stream tube theory

    International Nuclear Information System (INIS)

    Elmabrok, Ali Mohamed; Al-makhlufi, Ahmed A.

    2006-01-01

    The performance of the Darrieus vertical axis turbine is comparable with that of the more common horizontal axis machines. It has a number of aerodynamic and structural advantages over HAWTS. However the darrieus turbines are not self-starting at low wind speeds which is a considerable disadvantage for a simple small scale installation. Generally, papers concerning vertical axis turbine do not study the behavior of the rotor at low tip speed ratios. Therefore they do not deal with the self starting problems. A number of analytical methods were investigated to see whether they could predict the starting performance of vertical axis turbines. The chosen methods and 'actuator disc theory' for multiple stream tubes. In this paper the multiple stream tube model is applied using two discs in tandem. The computational analysis of all models simulates the blade aerodynamics throughout the full range of incidence from 180 degree centigrade. The effects of varying various geometric parameters of the windmill upon the performance of the rotor are investigated to find a design with improved self starting characteristics. The best agreement between theory and experiment was obtained using the multiple stream tube (double disc) models.(Author)

  13. Efficient Estimation of Dynamic Density Functions with Applications in Streaming Data

    KAUST Repository

    Qahtan, Abdulhakim Ali Ali

    2016-01-01

    application is to detect outliers in data streams from sensor networks based on the estimated PDF. The method detects outliers accurately and outperforms baseline methods designed for detecting and cleaning outliers in sensor data. The third application

  14. Predictive analytics on evolving data streams anticipating and adapting to changes in known and unknown contexts

    NARCIS (Netherlands)

    Pechenizkiy, M.

    2015-01-01

    Ever increasing volumes of sensor readings, transactional records, web data and event logs call for next generation of big data mining technology providing effective and efficient tools for making use of the streaming data. Predictive analytics on data streams is actively studied in research

  15. Association rule extraction from XML stream data for wireless sensor networks.

    Science.gov (United States)

    Paik, Juryon; Nam, Junghyun; Kim, Ung Mo; Won, Dongho

    2014-07-18

    With the advances of wireless sensor networks, they yield massive volumes of disparate, dynamic and geographically-distributed and heterogeneous data. The data mining community has attempted to extract knowledge from the huge amount of data that they generate. However, previous mining work in WSNs has focused on supporting simple relational data structures, like one table per network, while there is a need for more complex data structures. This deficiency motivates XML, which is the current de facto format for the data exchange and modeling of a wide variety of data sources over the web, to be used in WSNs in order to encourage the interchangeability of heterogeneous types of sensors and systems. However, mining XML data for WSNs has two challenging issues: one is the endless data flow; and the other is the complex tree structure. In this paper, we present several new definitions and techniques related to association rule mining over XML data streams in WSNs. To the best of our knowledge, this work provides the first approach to mining XML stream data that generates frequent tree items without any redundancy.

  16. Shifting stream planform state decreases stream productivity yet increases riparian animal production

    Science.gov (United States)

    Venarsky, Michael P.; Walters, David M.; Hall, Robert O.; Livers, Bridget; Wohl, Ellen

    2018-01-01

    In the Colorado Front Range (USA), disturbance history dictates stream planform. Undisturbed, old-growth streams have multiple channels and large amounts of wood and depositional habitat. Disturbed streams (wildfires and logging production, emerging aquatic insect flux, and riparian spider biomass. Organic matter and macroinvertebrate production did not differ among sites per unit area (m−2), but values were 2 ×–21 × higher in undisturbed reaches per unit of stream valley (m−1 valley) because total stream area was higher in undisturbed reaches. Insect emergence was similar among streams at the per unit area and per unit of stream valley. However, rescaling insect emergence to per meter of stream bank showed that the emerging insect biomass reaching the stream bank was lower in undisturbed sites because multi-channel reaches had 3 × more stream bank than single-channel reaches. Riparian spider biomass followed the same pattern as emerging aquatic insects, and we attribute this to bottom-up limitation caused by the multi-channeled undisturbed sites diluting prey quantity (emerging insects) reaching the stream bank (riparian spider habitat). These results show that historic landscape disturbances continue to influence stream and riparian communities in the Colorado Front Range. However, these legacy effects are only weakly influencing habitat-specific function and instead are primarily influencing stream–riparian community productivity by dictating both stream planform (total stream area, total stream bank length) and the proportional distribution of specific habitat types (pools vs riffles).

  17. On-stream chemical element monitor

    International Nuclear Information System (INIS)

    Averitt, O.R.; Dorsch, R.R.

    1979-01-01

    An apparatus and method for on-stream chemical element monitoring are described wherein a multiplicity of sample streams are flowed continuously through individual analytical cells and fluorescence analyses are performed on the sample streams in sequence, together with a method of controlling the time duration of each analysis as a function of the concomitant radiation exposure of a preselected perforate reference material interposed in the sample-radiation source path

  18. Vectorization for Molecular Dynamics on Intel Xeon Phi Corpocessors

    Science.gov (United States)

    Yi, Hongsuk

    2014-03-01

    Many modern processors are capable of exploiting data-level parallelism through the use of single instruction multiple data (SIMD) execution. The new Intel Xeon Phi coprocessor supports 512 bit vector registers for the high performance computing. In this paper, we have developed a hierarchical parallelization scheme for accelerated molecular dynamics simulations with the Terfoff potentials for covalent bond solid crystals on Intel Xeon Phi coprocessor systems. The scheme exploits multi-level parallelism computing. We combine thread-level parallelism using a tightly coupled thread-level and task-level parallelism with 512-bit vector register. The simulation results show that the parallel performance of SIMD implementations on Xeon Phi is apparently superior to their x86 CPU architecture.

  19. An Approximate L p Difference Algorithm for Massive Data Streams

    Directory of Open Access Journals (Sweden)

    Jessica H. Fong

    2001-12-01

    Full Text Available Several recent papers have shown how to approximate the difference ∑ i |a i-b i | or ∑|a i-b i | 2 between two functions, when the function values a i and b i are given in a data stream, and their order is chosen by an adversary. These algorithms use little space (much less than would be needed to store the entire stream and little time to process each item in the stream. They approximate with small relative error. Using different techniques, we show how to approximate the L p-difference ∑ i |a i-b i | p for any rational-valued p∈(0,2], with comparable efficiency and error. We also show how to approximate ∑ i |a i-b i | p for larger values of p but with a worse error guarantee. Our results fill in gaps left by recent work, by providing an algorithm that is precisely tunable for the application at hand. These results can be used to assess the difference between two chronologically or physically separated massive data sets, making one quick pass over each data set, without buffering the data or requiring the data source to pause. For example, one can use our techniques to judge whether the traffic on two remote network routers are similar without requiring either router to transmit a copy of its traffic. A web search engine could use such algorithms to construct a library of small ``sketches,'' one for each distinct page on the web; one can approximate the extent to which new web pages duplicate old ones by comparing the sketches of the web pages. Such techniques will become increasingly important as the enormous scale, distributional nature, and one-pass processing requirements of data sets become more commonplace.

  20. Implementation of a Multichannel Serial Data Streaming Algorithm using the Xilinx Serial RapidIO Solution

    Science.gov (United States)

    Doxley, Charles A.

    2016-01-01

    In the current world of applications that use reconfigurable technology implemented on field programmable gate arrays (FPGAs), there is a need for flexible architectures that can grow as the systems evolve. A project has limited resources and a fixed set of requirements that development efforts are tasked to meet. Designers must develop robust solutions that practically meet the current customer demands and also have the ability to grow for future performance. This paper describes the development of a high speed serial data streaming algorithm that allows for transmission of multiple data channels over a single serial link. The technique has the ability to change to meet new applications developed for future design considerations. This approach uses the Xilinx Serial RapidIO LOGICORE Solution to implement a flexible infrastructure to meet the current project requirements with the ability to adapt future system designs.

  1. Shifting stream planform state decreases stream productivity yet increases riparian animal production

    Science.gov (United States)

    Venarsky, Michael P.; Walters, David M.; Hall, Robert O.; Livers, Bridget; Wohl, Ellen

    2018-01-01

    In the Colorado Front Range (USA), disturbance history dictates stream planform. Undisturbed, old-growth streams have multiple channels and large amounts of wood and depositional habitat. Disturbed streams (wildfires and logging tested how these opposing stream states influenced organic matter, benthic macroinvertebrate secondary production, emerging aquatic insect flux, and riparian spider biomass. Organic matter and macroinvertebrate production did not differ among sites per unit area (m−2), but values were 2 ×–21 × higher in undisturbed reaches per unit of stream valley (m−1 valley) because total stream area was higher in undisturbed reaches. Insect emergence was similar among streams at the per unit area and per unit of stream valley. However, rescaling insect emergence to per meter of stream bank showed that the emerging insect biomass reaching the stream bank was lower in undisturbed sites because multi-channel reaches had 3 × more stream bank than single-channel reaches. Riparian spider biomass followed the same pattern as emerging aquatic insects, and we attribute this to bottom-up limitation caused by the multi-channeled undisturbed sites diluting prey quantity (emerging insects) reaching the stream bank (riparian spider habitat). These results show that historic landscape disturbances continue to influence stream and riparian communities in the Colorado Front Range. However, these legacy effects are only weakly influencing habitat-specific function and instead are primarily influencing stream–riparian community productivity by dictating both stream planform (total stream area, total stream bank length) and the proportional distribution of specific habitat types (pools vs riffles).

  2. Functional trait composition of aquatic plants can serve to disentangle multiple interacting stressors in lowland streams

    Energy Technology Data Exchange (ETDEWEB)

    Baattrup-Pedersen, Annette, E-mail: abp@bios.au.dk [Department of Bioscience, Aarhus University, Vejlsøvej 25, P.O. Box 314, DK-8600 Silkeborg (Denmark); Göthe, Emma [Department of Bioscience, Aarhus University, Vejlsøvej 25, P.O. Box 314, DK-8600 Silkeborg (Denmark); Riis, Tenna [Department of Bioscience, Aarhus University, Ole Worms Allé 1, Building 1135, Room 217, DK-8000 Aarhus C (Denmark); O' Hare, Matthew T. [Centre for Ecology and Hydrology, Bush Estate, Penicuik EH26 0QB (United Kingdom)

    2016-02-01

    Historically, close attention has been paid to negative impacts associated with nutrient loads to streams and rivers, but today hydromorphological alterations are considered increasingly implicated when lowland streams do not achieve good ecological status. Here, we explore if trait-abundance patterns of aquatic plants change along gradients in hydromorphological degradation and eutrophication in lowland stream sites located in Denmark. Specifically, we hypothesised that: i) changes in trait-abundance patterns occur along gradients in hydromorphological degradation and ii) trait-abundance patterns can serve to disentangle effects of eutrophication and hydromorphological degradation in lowland streams reflecting that the mechanisms behind changes differ. We used monitoring data from a total of 147 stream reaches with combined data on aquatic plant species abundance, catchment land use, hydromorphological alterations (i.e. planform, cross section, weed cutting) and water chemistry parameters. Traits related to life form, dispersal, reproduction and survival together with ecological preference values for nutrients and light (Ellenberg N and L) were allocated to 41 species representing 79% of the total species pool. We found clear evidence that habitat degradation (hydromorphological alterations and eutrophication) mediated selective changes in the trait-abundance patterns of the plant community. Specific traits could distinguish hydromorphological degradation (free-floating, surface; anchored floating leaves; anchored heterophylly) from eutrophication (free-floating, submerged; leaf area). We provide a conceptual framework for interpretation of how eutrophication and hydromorphological degradation interact and how this is reflected in trait-abundance patterns in aquatic plant communities in lowland streams. Our findings support the merit of trait-based approaches in biomonitoring as they shed light on mechanisms controlling structural changes under environmental

  3. Functional trait composition of aquatic plants can serve to disentangle multiple interacting stressors in lowland streams

    International Nuclear Information System (INIS)

    Baattrup-Pedersen, Annette; Göthe, Emma; Riis, Tenna; O'Hare, Matthew T.

    2016-01-01

    Historically, close attention has been paid to negative impacts associated with nutrient loads to streams and rivers, but today hydromorphological alterations are considered increasingly implicated when lowland streams do not achieve good ecological status. Here, we explore if trait-abundance patterns of aquatic plants change along gradients in hydromorphological degradation and eutrophication in lowland stream sites located in Denmark. Specifically, we hypothesised that: i) changes in trait-abundance patterns occur along gradients in hydromorphological degradation and ii) trait-abundance patterns can serve to disentangle effects of eutrophication and hydromorphological degradation in lowland streams reflecting that the mechanisms behind changes differ. We used monitoring data from a total of 147 stream reaches with combined data on aquatic plant species abundance, catchment land use, hydromorphological alterations (i.e. planform, cross section, weed cutting) and water chemistry parameters. Traits related to life form, dispersal, reproduction and survival together with ecological preference values for nutrients and light (Ellenberg N and L) were allocated to 41 species representing 79% of the total species pool. We found clear evidence that habitat degradation (hydromorphological alterations and eutrophication) mediated selective changes in the trait-abundance patterns of the plant community. Specific traits could distinguish hydromorphological degradation (free-floating, surface; anchored floating leaves; anchored heterophylly) from eutrophication (free-floating, submerged; leaf area). We provide a conceptual framework for interpretation of how eutrophication and hydromorphological degradation interact and how this is reflected in trait-abundance patterns in aquatic plant communities in lowland streams. Our findings support the merit of trait-based approaches in biomonitoring as they shed light on mechanisms controlling structural changes under environmental

  4. Geospatial data stream processing in Python using FOSS4G components

    CSIR Research Space (South Africa)

    McFerren, Graeme

    2016-07-01

    Full Text Available that Swordfish is stable for long running applications, though its early deployment in particular use cases suggest it is reasonably stable. Swordfish is currently limited to holding state in memory; further work may be necessary to develop... mechanisms to serialise state, especially in use cases where recovery of the streaming topology state may be necessary. A long term view of Swordfish development is the provision of a streaming data management platform, (as is provided by ESRI Geo...

  5. MonetDB/DataCell: Online Analytics in a Streaming Column-Store

    NARCIS (Netherlands)

    E. Liarou (Erietta); S. Idreos (Stratos); S. Manegold (Stefan); M.L. Kersten (Martin)

    2012-01-01

    textabstractIn DataCell, we design streaming functionalities in a mod- ern relational database kernel which targets big data analyt- ics. This includes exploitation of both its storage/execution engine and its optimizer infrastructure. We investigate the opportunities and challenges that arise with

  6. MonetDB/DataCell: online analytics in a streaming column-store

    NARCIS (Netherlands)

    Liarou, E.; Idreos, S.; Manegold, S.; Kersten, M.

    2012-01-01

    In DataCell, we design streaming functionalities in a modern relational database kernel which targets big data analytics. This includes exploitation of both its storage/execution engine and its optimizer infrastructure. We investigate the opportunities and challenges that arise with such a direction

  7. Towards the Development of a Taxonomy for Visualisation of Streamed Geospatial Data

    Science.gov (United States)

    Sibolla, B. H.; Van Zyl, T.; Coetzee, S.

    2016-06-01

    Geospatial data has very specific characteristics that need to be carefully captured in its visualisation, in order for the user and the viewer to gain knowledge from it. The science of visualisation has gained much traction over the last decade as a response to various visualisation challenges. During the development of an open source based, dynamic two-dimensional visualisation library, that caters for geospatial streaming data, it was found necessary to conduct a review of existing geospatial visualisation taxonomies. The review was done in order to inform the design phase of the library development, such that either an existing taxonomy can be adopted or extended to fit the needs at hand. The major challenge in this case is to develop dynamic two dimensional visualisations that enable human interaction in order to assist the user to understand the data streams that are continuously being updated. This paper reviews the existing geospatial data visualisation taxonomies that have been developed over the years. Based on the review, an adopted taxonomy for visualisation of geospatial streaming data is presented. Example applications of this taxonomy are also provided. The adopted taxonomy will then be used to develop the information model for the visualisation library in a further study.

  8. Pilot-Streaming: Design Considerations for a Stream Processing Framework for High-Performance Computing

    OpenAIRE

    Andre Luckow; Peter Kasson; Shantenu Jha

    2016-01-01

    This White Paper (submitted to STREAM 2016) identifies an approach to integrate streaming data with HPC resources. The paper outlines the design of Pilot-Streaming, which extends the concept of Pilot-abstraction to streaming real-time data.

  9. Incremental temporal pattern mining using efficient batch-free stream clustering

    NARCIS (Netherlands)

    Lu, Y.; Hassani, M.; Seidl, T.

    2017-01-01

    This paper address the problem of temporal pattern mining from multiple data streams containing temporal events. Temporal events are considered as real world events aligned with comprehensive starting and ending timing information rather than simple integer timestamps. Predefined relations, such as

  10. Association Rule Extraction from XML Stream Data for Wireless Sensor Networks

    Science.gov (United States)

    Paik, Juryon; Nam, Junghyun; Kim, Ung Mo; Won, Dongho

    2014-01-01

    With the advances of wireless sensor networks, they yield massive volumes of disparate, dynamic and geographically-distributed and heterogeneous data. The data mining community has attempted to extract knowledge from the huge amount of data that they generate. However, previous mining work in WSNs has focused on supporting simple relational data structures, like one table per network, while there is a need for more complex data structures. This deficiency motivates XML, which is the current de facto format for the data exchange and modeling of a wide variety of data sources over the web, to be used in WSNs in order to encourage the interchangeability of heterogeneous types of sensors and systems. However, mining XML data for WSNs has two challenging issues: one is the endless data flow; and the other is the complex tree structure. In this paper, we present several new definitions and techniques related to association rule mining over XML data streams in WSNs. To the best of our knowledge, this work provides the first approach to mining XML stream data that generates frequent tree items without any redundancy. PMID:25046017

  11. A catchment scale evaluation of multiple stressor effects in headwater streams

    DEFF Research Database (Denmark)

    Rasmussen, J. J.; McKnight, Ursula S.; Loinaz, Maria Christina

    2013-01-01

    studied 11 headwater streams in the Hove catchment in the Copenhagen region. All sites had substantial physical habitat and water quality impairments due to anthropogenic influence (intensive agriculture, urban settlements, contaminated sites and low base-flow due to water abstraction activities...... insecticides were probably essential contributors to the overall ecological impairment of these streams. Our results suggest that headwater streams should be considered in future management and mitigation plans. Catchment-based management is necessary because several anthropogenic stressors exceeded...

  12. Hydrogeochemical and stream sediment reconnaissance basic data for Lawton NTMS quadrangle, Oklahoma; Texas

    International Nuclear Information System (INIS)

    1978-01-01

    Field and laboratory data are presented for 703 groundwater and 782 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. Groundwater data indicate that the most promising areas for potential uranium mineralization occur in the Lower Permian units surrounding the granite outcrops of the Wichita Mountains. Waters from the Hennessey and Clearfork Groups and the Garber Sandstone contain the highest uranium values. Elements associated with the uranium are arsenic, boron, barium, molybdenum, sodium, selenium, and vandium. Stream sediment data indicate that the promising areas for potential uranium mineralization occur around the Wichita Mountains where stream sediments are derived from the Lower Permian Post Oak Conglomerate, Hennessey Group, and Garber Sandstone and from the Cambrian igneous rocks. Other areas of interest occur (1) in the western part of the quadrangle where the sediments are derived from rocks of the El Reno Group, and (2) along the southern border of the quadrangle where the sediments are derived from the Wichita Group

  13. Truck Route Choice Modeling using Large Streams of GPS Data

    Science.gov (United States)

    2017-07-31

    The primary goal of this research was to use large streams of truck-GPS data to analyze travel routes (or paths) chosen by freight trucks to travel between different origin and destination (OD) location pairs in metropolitan regions of Florida. Two s...

  14. Dynamically Scaling Apache Storm for the Analysis of Streaming Data

    NARCIS (Netherlands)

    Veen, J.S. van der; Waaij, B.D. van der; Lazovik, E.; Wijbrandi, W.E.; Meijer, R.J.

    2015-01-01

    Stream processing platforms allow applications to analyse incoming data continuously. Several use cases exist that make use of these capabilities, ranging from monitoring of physical infrastructures to pre selecting video surveillance feeds for human inspection. It is difficult to predict how much

  15. Assessing the impact of groundwater contamination on stream water quality by multiple approaches at the groundwater-surface water interface (Invited Presentation)

    DEFF Research Database (Denmark)

    Bjerg, Poul Løgstrup; Rønde, Vinni Kampman; Balbarini, Nicola

    Contaminants such as chlorinated solvents and pesticides, as well as new classes of compounds or emerging micropollutants are extensively produced, utilized and then discarded in society and subsequently released to streams from multiple point and diffuse sources. Sustainable management of water...

  16. Design of Optimized Multimedia Data Streaming Management Using OMDSM over Mobile Networks

    Directory of Open Access Journals (Sweden)

    Byungjoo Park

    2017-01-01

    Full Text Available Mobility management is an essential challenge for supporting reliable multimedia data streaming over wireless and mobile networks in the Internet of Things (IoT for location-based mobile marketing applications. The communications among mobile nodes for IoT need to have a seamless handover for delivering high quality multimedia services. The Internet Engineering Task Force (IETF mobility management schemes are the proposals for handling the routing of IPv6 packets to mobile nodes that have moved away from their home network. However, the standard mobility management scheme cannot prevent packet losses due to longer handover latency. In this article, a new enhanced data streaming route optimization scheme is introduced that uses an optimized Transmission Control Protocol (TCP realignment algorithm in order to prevent the packet disordering problem whenever the nodes in the IoT environment are communicating with each other. With the proposed scheme, data packets sequence realignment can be prevented, the packet traffic speed can be controlled, and the TCP performance can be improved. The experimental results show that managing the packet order in proposed new scheme remarkably increases the overall TCP performance over mobile networks within the IoT environment thus ensuring the high quality of service (QoS for multimedia data streaming in location-based mobile marketing applications.

  17. Relative performance of three stream bed stability indices as indicators of stream health.

    Science.gov (United States)

    Kusnierz, Paul C; Holbrook, Christopher M

    2017-10-16

    Bed stability is an important stream habitat attribute because it affects geomorphology and biotic communities. Natural resource managers desire indices of bed stability that can be used under a wide range of geomorphic conditions, are biologically meaningful, and are easily incorporated into sampling protocols. To eliminate potential bias due to presence of instream wood and increase precision of stability values, we modified a stream bed instability index (ISI) to include measurements of bankfull depth (d bf ) and median particle diameter (D 50 ) only in riffles and increased the pebble count to decrease variability (i.e., increase precision) in D 50 . The new riffle-based instability index (RISI) was compared to two established indices: ISI and the riffle stability index (RSI). RISI and ISI were strongly associated with each other but neither was closely associated with RSI. RISI and ISI were closely associated with both a diatom- and two macrovertebrate-based stream health indices, but RSI was only weakly associated with the macroinvertebrate indices. Unexpectedly, precision of D 50 did not differ between RISI and ISI. Results suggest that RISI is a viable alternative to both ISI and RSI for evaluating bed stability in multiple stream types. With few data requirements and a simple protocol, RISI may also better conform to riffle-based sampling methods used by some water quality practitioners.

  18. The influence of data characteristics on detecting wetland/stream surface-water connections in the Delmarva Peninsula, Maryland and Delaware

    Science.gov (United States)

    Vanderhoof, Melanie; Distler, Hayley; Lang, Megan W.; Alexander, Laurie C.

    2018-01-01

    The dependence of downstream waters on upstream ecosystems necessitates an improved understanding of watershed-scale hydrological interactions including connections between wetlands and streams. An evaluation of such connections is challenging when, (1) accurate and complete datasets of wetland and stream locations are often not available and (2) natural variability in surface-water extent influences the frequency and duration of wetland/stream connectivity. The Upper Choptank River watershed on the Delmarva Peninsula in eastern Maryland and Delaware is dominated by a high density of small, forested wetlands. In this analysis, wetland/stream surface water connections were quantified using multiple wetland and stream datasets, including headwater streams and depressions mapped from a lidar-derived digital elevation model. Surface-water extent was mapped across the watershed for spring 2015 using Landsat-8, Radarsat-2 and Worldview-3 imagery. The frequency of wetland/stream connections increased as a more complete and accurate stream dataset was used and surface-water extent was included, in particular when the spatial resolution of the imagery was finer (i.e., watershed contributing direct surface water runoff to streamflow. This finding suggests that our interpretation of the frequency and duration of wetland/stream connections will be influenced not only by the spatial and temporal characteristics of wetlands, streams and potential flowpaths, but also by the completeness, accuracy and resolution of input datasets.

  19. STREAM2016: Streaming Requirements, Experience, Applications and Middleware Workshop

    Energy Technology Data Exchange (ETDEWEB)

    Fox, Geoffrey [Indiana Univ., Bloomington, IN (United States); Jha, Shantenu [Rutgers Univ., New Brunswick, NJ (United States); Ramakrishnan, Lavanya [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)

    2016-10-01

    The Department of Energy (DOE) Office of Science (SC) facilities including accelerators, light sources and neutron sources and sensors that study, the environment, and the atmosphere, are producing streaming data that needs to be analyzed for next-generation scientific discoveries. There has been an explosion of new research and technologies for stream analytics arising from the academic and private sectors. However, there has been no corresponding effort in either documenting the critical research opportunities or building a community that can create and foster productive collaborations. The two-part workshop series, STREAM: Streaming Requirements, Experience, Applications and Middleware Workshop (STREAM2015 and STREAM2016), were conducted to bring the community together and identify gaps and future efforts needed by both NSF and DOE. This report describes the discussions, outcomes and conclusions from STREAM2016: Streaming Requirements, Experience, Applications and Middleware Workshop, the second of these workshops held on March 22-23, 2016 in Tysons, VA. STREAM2016 focused on the Department of Energy (DOE) applications, computational and experimental facilities, as well software systems. Thus, the role of “streaming and steering” as a critical mode of connecting the experimental and computing facilities was pervasive through the workshop. Given the overlap in interests and challenges with industry, the workshop had significant presence from several innovative companies and major contributors. The requirements that drive the proposed research directions, identified in this report, show an important opportunity for building competitive research and development program around streaming data. These findings and recommendations are consistent with vision outlined in NRC Frontiers of Data and National Strategic Computing Initiative (NCSI) [1, 2]. The discussions from the workshop are captured as topic areas covered in this report's sections. The report

  20. Simultaneous heat integration and techno-economic optimization of Organic Rankine Cycle (ORC) for multiple waste heat stream recovery

    International Nuclear Information System (INIS)

    Yu, Haoshui; Eason, John; Biegler, Lorenz T.; Feng, Xiao

    2017-01-01

    In the past decades, the Organic Rankine Cycle (ORC) has become a promising technology for low and medium temperature energy utilization. In refineries, there are usually multiple waste heat streams to be recovered. From a safety and controllability perspective, using an intermedium (hot water) to recover waste heat before releasing heat to the ORC system is more favorable than direct integration. The mass flowrate of the intermediate hot water stream determines the amount of waste heat recovered and the final hot water temperature affects the thermal efficiency of ORC. Both, in turn, exert great influence on the power output. Therefore, the hot water mass flowrate is a critical decision variable for the optimal design of the system. This study develops a model for techno-economic optimization of an ORC with simultaneous heat recovery and capital cost optimization. The ORC is modeled using rigorous thermodynamics with the concept of state points. The task of waste heat recovery using the hot water intermedium is modeled using the Duran-Grossmann model for simultaneous heat integration and process optimization. The combined model determines the optimal design of an ORC that recovers multiple waste heat streams in a large scale background process using an intermediate heat transfer stream. In particular, the model determines the optimal heat recovery approach temperature (HRAT), the utility load of the background process, and the optimal operating conditions of the ORC simultaneously. The effectiveness of this method is demonstrated with a case study that uses a refinery as the background process. Sensitivity of the optimal solution to the parameters (electricity price, utility cost) is quantified in this paper. - Highlights: • A new model for Organic Rankine cycle design optimization is presented. • Process heat integration and ORC are considered simultaneously. • Rigorous equation oriented models of the ORC are used for accurate results. • Impact of working

  1. Deploy Nalu/Kokkos algorithmic infrastructure with performance benchmarking.

    Energy Technology Data Exchange (ETDEWEB)

    Domino, Stefan P. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Ananthan, Shreyas [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Knaus, Robert C. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Williams, Alan B. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-09-29

    The former Nalu interior heterogeneous algorithm design, which was originally designed to manage matrix assembly operations over all elemental topology types, has been modified to operate over homogeneous collections of mesh entities. This newly templated kernel design allows for removal of workset variable resize operations that were formerly required at each loop over a Sierra ToolKit (STK) bucket (nominally, 512 entities in size). Extensive usage of the Standard Template Library (STL) std::vector has been removed in favor of intrinsic Kokkos memory views. In this milestone effort, the transition to Kokkos as the underlying infrastructure to support performance and portability on many-core architectures has been deployed for key matrix algorithmic kernels. A unit-test driven design effort has developed a homogeneous entity algorithm that employs a team-based thread parallelism construct. The STK Single Instruction Multiple Data (SIMD) infrastructure is used to interleave data for improved vectorization. The collective algorithm design, which allows for concurrent threading and SIMD management, has been deployed for the core low-Mach element- based algorithm. Several tests to ascertain SIMD performance on Intel KNL and Haswell architectures have been carried out. The performance test matrix includes evaluation of both low- and higher-order methods. The higher-order low-Mach methodology builds on polynomial promotion of the core low-order control volume nite element method (CVFEM). Performance testing of the Kokkos-view/SIMD design indicates low-order matrix assembly kernel speed-up ranging between two and four times depending on mesh loading and node count. Better speedups are observed for higher-order meshes (currently only P=2 has been tested) especially on KNL. The increased workload per element on higher-order meshes bene ts from the wide SIMD width on KNL machines. Combining multiple threads with SIMD on KNL achieves a 4.6x speedup over the baseline, with

  2. Hydrogeochemical and stream sediment reconnaissance basic data for Dickinson NTMS Quadrangle, North Dakota

    International Nuclear Information System (INIS)

    1980-01-01

    Results of a reconnaissance geochemical survey of the Dickinson Quadrangle, North Dakota are reported. Field and laboratory data are presented for 544 groundwater and 554 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided, and pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Interpretation of the groundwater data indicates that scattered localities in the central portion of the quadrangle appear most promising for uranium mineralization. High values of uranium in this area are usually found in waters of the Sentinel Butte and Tongue River Formations. Uranium is believed to be concentrated in the lignite beds of the Fort Union Group, with concentrations increasing with proximity to the pre-Oligocene unconformity. Stream sediment data indicate high uranium values distributed over the central area of the quadrangle. Uranium in stream sediments does not appear to be associated with any particular geologic unit and is perhaps following a structural trend

  3. Autonomic intrusion detection: Adaptively detecting anomalies over unlabeled audit data streams in computer networks

    KAUST Repository

    Wang, Wei; Guyet, Thomas; Quiniou, René ; Cordier, Marie-Odile; Masseglia, Florent; Zhang, Xiangliang

    2014-01-01

    In this work, we propose a novel framework of autonomic intrusion detection that fulfills online and adaptive intrusion detection over unlabeled HTTP traffic streams in computer networks. The framework holds potential for self-managing: self-labeling, self-updating and self-adapting. Our framework employs the Affinity Propagation (AP) algorithm to learn a subject’s behaviors through dynamical clustering of the streaming data. It automatically labels the data and adapts to normal behavior changes while identifies anomalies. Two large real HTTP traffic streams collected in our institute as well as a set of benchmark KDD’99 data are used to validate the framework and the method. The test results show that the autonomic model achieves better results in terms of effectiveness and efficiency compared to adaptive Sequential Karhunen–Loeve method and static AP as well as three other static anomaly detection methods, namely, k-NN, PCA and SVM.

  4. Autonomic intrusion detection: Adaptively detecting anomalies over unlabeled audit data streams in computer networks

    KAUST Repository

    Wang, Wei

    2014-06-22

    In this work, we propose a novel framework of autonomic intrusion detection that fulfills online and adaptive intrusion detection over unlabeled HTTP traffic streams in computer networks. The framework holds potential for self-managing: self-labeling, self-updating and self-adapting. Our framework employs the Affinity Propagation (AP) algorithm to learn a subject’s behaviors through dynamical clustering of the streaming data. It automatically labels the data and adapts to normal behavior changes while identifies anomalies. Two large real HTTP traffic streams collected in our institute as well as a set of benchmark KDD’99 data are used to validate the framework and the method. The test results show that the autonomic model achieves better results in terms of effectiveness and efficiency compared to adaptive Sequential Karhunen–Loeve method and static AP as well as three other static anomaly detection methods, namely, k-NN, PCA and SVM.

  5. Towards a streaming model for nested data parallelism

    DEFF Research Database (Denmark)

    Madsen, Frederik Meisner; Filinski, Andrzej

    2013-01-01

    The language-integrated cost semantics for nested data parallelism pioneered by NESL provides an intuitive, high-level model for predicting performance and scalability of parallel algorithms with reasonable accuracy. However, this predictability, obtained through a uniform, parallelism-flattening......The language-integrated cost semantics for nested data parallelism pioneered by NESL provides an intuitive, high-level model for predicting performance and scalability of parallel algorithms with reasonable accuracy. However, this predictability, obtained through a uniform, parallelism......-processable in a streaming fashion. This semantics is directly compatible with previously proposed piecewise execution models for nested data parallelism, but allows the expected space usage to be reasoned about directly at the source-language level. The language definition and implementation are still very much work...

  6. Gesture Recognition from Data Streams of Human Motion Sensor Using Accelerated PSO Swarm Search Feature Selection Algorithm

    Directory of Open Access Journals (Sweden)

    Simon Fong

    2015-01-01

    Full Text Available Human motion sensing technology gains tremendous popularity nowadays with practical applications such as video surveillance for security, hand signing, and smart-home and gaming. These applications capture human motions in real-time from video sensors, the data patterns are nonstationary and ever changing. While the hardware technology of such motion sensing devices as well as their data collection process become relatively mature, the computational challenge lies in the real-time analysis of these live feeds. In this paper we argue that traditional data mining methods run short of accurately analyzing the human activity patterns from the sensor data stream. The shortcoming is due to the algorithmic design which is not adaptive to the dynamic changes in the dynamic gesture motions. The successor of these algorithms which is known as data stream mining is evaluated versus traditional data mining, through a case of gesture recognition over motion data by using Microsoft Kinect sensors. Three different subjects were asked to read three comic strips and to tell the stories in front of the sensor. The data stream contains coordinates of articulation points and various positions of the parts of the human body corresponding to the actions that the user performs. In particular, a novel technique of feature selection using swarm search and accelerated PSO is proposed for enabling fast preprocessing for inducing an improved classification model in real-time. Superior result is shown in the experiment that runs on this empirical data stream. The contribution of this paper is on a comparative study between using traditional and data stream mining algorithms and incorporation of the novel improved feature selection technique with a scenario where different gesture patterns are to be recognized from streaming sensor data.

  7. Performance Modeling in CUDA Streams - A Means for High-Throughput Data Processing.

    Science.gov (United States)

    Li, Hao; Yu, Di; Kumar, Anand; Tu, Yi-Cheng

    2014-10-01

    Push-based database management system (DBMS) is a new type of data processing software that streams large volume of data to concurrent query operators. The high data rate of such systems requires large computing power provided by the query engine. In our previous work, we built a push-based DBMS named G-SDMS to harness the unrivaled computational capabilities of modern GPUs. A major design goal of G-SDMS is to support concurrent processing of heterogenous query processing operations and enable resource allocation among such operations. Understanding the performance of operations as a result of resource consumption is thus a premise in the design of G-SDMS. With NVIDIA's CUDA framework as the system implementation platform, we present our recent work on performance modeling of CUDA kernels running concurrently under a runtime mechanism named CUDA stream . Specifically, we explore the connection between performance and resource occupancy of compute-bound kernels and develop a model that can predict the performance of such kernels. Furthermore, we provide an in-depth anatomy of the CUDA stream mechanism and summarize the main kernel scheduling disciplines in it. Our models and derived scheduling disciplines are verified by extensive experiments using synthetic and real-world CUDA kernels.

  8. Troubled Waters: where Multiple Streams of Inequality Converge in the Math and Science Experiences of Nonprivileged Girls

    Science.gov (United States)

    Parrott, Laurel; Spatig, Linda; Kusimo, Patricia S.; Carter, Carolyn C.; Keyes, Marian

    Water is often hardest to navigate at the confluence of individual streams. As they experience math and science, nonprivileged girls maneuver through roiling waters where the streams of gender, ethnicity, poverty, place, and teaching practices converge. Just as waters of separate streams blend, these issues - too often considered separate factors - become blended and difficult to isolate, and the resulting turbulence produces a bumpy ride. We draw on 3 years of qualitative data collected as part of an intervention program to explore the math and science experiences and perceptions of a group of ethnically diverse, low socioeconomic status rural and urban adolescent Appalachian girls. After describing program and community contexts, we explore "opportunity to leant" issues - specifically, expectations, access to content, and support networks - and examine their schooling experiences against visions of science and math reform and pressures for accountability. Data are discussed within a framework of critical educational theory.

  9. Mahanaxar: quality of service guarantees in high-bandwidth, real-time streaming data storage

    Energy Technology Data Exchange (ETDEWEB)

    Bigelow, David [Los Alamos National Laboratory; Bent, John [Los Alamos National Laboratory; Chen, Hsing-Bung [Los Alamos National Laboratory; Brandt, Scott [UCSC

    2010-04-05

    Large radio telescopes, cyber-security systems monitoring real-time network traffic, and others have specialized data storage needs: guaranteed capture of an ultra-high-bandwidth data stream, retention of the data long enough to determine what is 'interesting,' retention of interesting data indefinitely, and concurrent read/write access to determine what data is interesting, without interrupting the ongoing capture of incoming data. Mahanaxar addresses this problem. Mahanaxar guarantees streaming real-time data capture at (nearly) the full rate of the raw device, allows concurrent read and write access to the device on a best-effort basis without interrupting the data capture, and retains data as long as possible given the available storage. It has built in mechanisms for reliability and indexing, can scale to meet arbitrary bandwidth requirements, and handles both small and large data elements equally well. Results from our prototype implementation shows that Mahanaxar provides both better guarantees and better performance than traditional file systems.

  10. 2008-09 National Rivers and Streams Assessment Fish Tissue Data Dictionary

    Science.gov (United States)

    The Office of Science and Technology (OST) is providing the fish tissue results from the 2008-09 National Rivers and Streams Assessment (NRSA). This document includes the “data dictionary” for Mercury, Selenium, PBDEs, PCBs, Pesticides and PFCs.

  11. A reference web architecture and patterns for real-time visual analytics on large streaming data

    Science.gov (United States)

    Kandogan, Eser; Soroker, Danny; Rohall, Steven; Bak, Peter; van Ham, Frank; Lu, Jie; Ship, Harold-Jeffrey; Wang, Chun-Fu; Lai, Jennifer

    2013-12-01

    Monitoring and analysis of streaming data, such as social media, sensors, and news feeds, has become increasingly important for business and government. The volume and velocity of incoming data are key challenges. To effectively support monitoring and analysis, statistical and visual analytics techniques need to be seamlessly integrated; analytic techniques for a variety of data types (e.g., text, numerical) and scope (e.g., incremental, rolling-window, global) must be properly accommodated; interaction, collaboration, and coordination among several visualizations must be supported in an efficient manner; and the system should support the use of different analytics techniques in a pluggable manner. Especially in web-based environments, these requirements pose restrictions on the basic visual analytics architecture for streaming data. In this paper we report on our experience of building a reference web architecture for real-time visual analytics of streaming data, identify and discuss architectural patterns that address these challenges, and report on applying the reference architecture for real-time Twitter monitoring and analysis.

  12. Knoxville 10 x 20 NTMS area, North Carolina, South Carolina, and Tennessee: data release. Hydrogeochemical and stream sediment reconnaissance

    International Nuclear Information System (INIS)

    Baucom, E.I.; Ferguson, R.B.

    1979-05-01

    Stream sediment and stream water samples were collected from small streams at 1430 sites or at a nominal density of one site per 14 square kilometers (five square miles) in rural areas. Ground water samples were collected at 791 sites or at a nominal density of one site per 25 square kilometers (ten square miles). Neutron activation analysis (NAA) results are given for uranium and 16 other elements in sediments, and for uranium and 9 other elements in ground water. Key data from ground water sites include (1) water chemistry measurements (pH, conductivity, and alkalinity), (2) well depth, (3) elemental analyses (U, Br, Cl, F, Mg, Mn, Na, and V). Supplementary data include site descriptors (well age, frequency of use of well, etc.) and tabulated analytical data for Al and Dy. Key data from stream sediment sites include (1) water quality measurements (pH, conductivity, and alkalinity), and (2) important elemental analyses (U, Th, Hf, Al, Ce, Fe, Mn, Na, Sc, Ti, and V). Supplementary data from stream sediment sites include sample site descriptors (stream characteristics, vegetation, etc.) and additional elemental analyses

  13. A programming framework for data streaming on the Xeon Phi

    Science.gov (United States)

    Chapeland, S.; ALICE Collaboration

    2017-10-01

    ALICE (A Large Ion Collider Experiment) is the dedicated heavy-ion detector studying the physics of strongly interacting matter and the quark-gluon plasma at the CERN LHC (Large Hadron Collider). After the second long shut-down of the LHC, the ALICE detector will be upgraded to cope with an interaction rate of 50 kHz in Pb-Pb collisions, producing in the online computing system (O2) a sustained throughput of 3.4 TB/s. This data will be processed on the fly so that the stream to permanent storage does not exceed 90 GB/s peak, the raw data being discarded. In the context of assessing different computing platforms for the O2 system, we have developed a framework for the Intel Xeon Phi processors (MIC). It provides the components to build a processing pipeline streaming the data from the PC memory to a pool of permanent threads running on the MIC, and back to the host after processing. It is based on explicit offloading mechanisms (data transfer, asynchronous tasks) and basic building blocks (FIFOs, memory pools, C++11 threads). The user only needs to implement the processing method to be run on the MIC. We present in this paper the architecture, implementation, and performance of this system.

  14. Effects of forest harvesting on summer stream temperatures in New Brunswick, Canada: an inter-catchment, multiple-year comparison

    Directory of Open Access Journals (Sweden)

    C. P.-A. Bourque

    2001-01-01

    Full Text Available This paper presents a pre- and post-harvest comparison of stream temperatures collected in five neighbouring streams (sub-catchments over a period of five years (1994-1998. The aim of the study was to determine whether land cover changes from clear cutting in areas outside forest buffer zones (applied to streams >0.5 m wide might contribute to an increase in summer mean stream temperatures in buffered streams downslope by infusion of warmed surface and sub-surface water into the streams. Specific relationships were observed in all five forest streams investigated. To assist in the analysis, several spatially-relevant variables, such as land cover change, mid-summer potential solar radiation, flow accumulation, stream location and slope of the land were determined, in part, from existing aerial photographs, GIS-archived forest inventory data and a digital terrain model of the study area. Spatial calculations of insolation levels for July 15th were used as an index of mid-summer solar heating across sub-catchments. Analysis indicated that prior to the 1995 harvest, differences in stream temperature could be attributed to (i topographic position and catchment-to-sun orientation, (ii the level of cutting that occurred in the upper catchment prior to the start of the study, and (iii the average slope within harvested areas. Compared to the pre-harvest mean stream temperatures in 1994, mean temperatures in the three streams downslope from the 1995 harvest areas increased by 0.3 to 0.7°C (representing a 4-8% increase; p-value of normalised temperatures Keywords: terrain attributes, solar radiation, land cover, forest buffers, New Brunswick regulations, spatial modelling, DEM, forest covertypes

  15. Stream processing health card application.

    Science.gov (United States)

    Polat, Seda; Gündem, Taflan Imre

    2012-10-01

    In this paper, we propose a data stream management system embedded to a smart card for handling and storing user specific summaries of streaming data coming from medical sensor measurements and/or other medical measurements. The data stream management system that we propose for a health card can handle the stream data rates of commonly known medical devices and sensors. It incorporates a type of context awareness feature that acts according to user specific information. The proposed system is cheap and provides security for private data by enhancing the capabilities of smart health cards. The stream data management system is tested on a real smart card using both synthetic and real data.

  16. No double-dissociation between optic ataxia and visual agnosia: Multiple sub-streams for multiple visuo-manual integrations

    NARCIS (Netherlands)

    Pisella, L.; Binkofski, F.; Lasek, K.; Toni, I.; Rossetti, Y.

    2006-01-01

    The current dominant view of the visual system is marked by the functional and anatomical dissociation between a ventral stream specialised for perception and a dorsal stream specialised for action. The "double-dissociation" between visual agnosia (VA), a deficit of visual recognition, and optic

  17. The Open Source DataTurbine Initiative: Streaming Data Middleware for Environmental Observing Systems

    Science.gov (United States)

    Fountain T.; Tilak, S.; Shin, P.; Hubbard, P.; Freudinger, L.

    2009-01-01

    The Open Source DataTurbine Initiative is an international community of scientists and engineers sharing a common interest in real-time streaming data middleware and applications. The technology base of the OSDT Initiative is the DataTurbine open source middleware. Key applications of DataTurbine include coral reef monitoring, lake monitoring and limnology, biodiversity and animal tracking, structural health monitoring and earthquake engineering, airborne environmental monitoring, and environmental sustainability. DataTurbine software emerged as a commercial product in the 1990 s from collaborations between NASA and private industry. In October 2007, a grant from the USA National Science Foundation (NSF) Office of Cyberinfrastructure allowed us to transition DataTurbine from a proprietary software product into an open source software initiative. This paper describes the DataTurbine software and highlights key applications in environmental monitoring.

  18. Scalable Stream Coding for Adaptive Foveation Enhanced Percept Multimedia Information Communication for Interactive Medical Applications

    National Research Council Canada - National Science Library

    Khan, Javed

    2003-01-01

    .... The demonstrated systems include interactive perceptual transcoding where real-time eye-tracker data fuses with a passing stream, the active subnet diffusion coding-- where multiple active nodes...

  19. Construction of a digital elevation model: methods and parallelization

    International Nuclear Information System (INIS)

    Mazzoni, Christophe

    1995-01-01

    The aim of this work is to reduce the computation time needed to produce the Digital Elevation Models (DEM) by using a parallel machine. It is made in collaboration between the French 'Institut Geographique National' (IGN) and the Laboratoire d'Electronique de Technologie et d'Instrumentation (LETI) of the French Atomic Energy Commission (CEA). The IGN has developed a system which provides DEM that is used to produce topographic maps. The kernel of this system is the correlator, a software which automatically matches pairs of homologous points of a stereo-pair of photographs. Nevertheless the correlator is expensive In computing time. In order to reduce computation time and to produce the DEM with same accuracy that the actual system, we have parallelized the IGN's correlator on the OPENVISION system. This hardware solution uses a SIMD (Single Instruction Multiple Data) parallel machine SYMPATI-2, developed by the LETI that is involved in parallel architecture and image processing. Our analysis of the implementation has demonstrated the difficulty of efficient coupling between scalar and parallel structure. So we propose solutions to reinforce this coupling. In order to accelerate more the processing we evaluate SYMPHONIE, a SIMD calculator, successor of SYMPATI-2. On an other hand, we developed a multi-agent approach for what a MIMD (Multiple Instruction, Multiple Data) architecture is available. At last, we describe a Multi-SIMD architecture that conciliates our two approaches. This architecture offers a capacity to apprehend efficiently multi-level treatment image. It is flexible by its modularity, and its communication network supplies reliability that interest sensible systems. (author) [fr

  20. Baseline Channel Geometry and Aquatic Habitat Data for Selected Streams in the Matanuska-Susitna Valley, Alaska

    Science.gov (United States)

    Curran, Janet H.; Rice, William J.

    2009-01-01

    Small streams in the rapidly developing Matanuska-Susitna Valley in south-central Alaska are known to support anadromous and resident fish but little is known about their hydrologic and riparian conditions, or their sensitivity to the rapid development of the area or climate variability. To help address this need, channel geometry and aquatic habitat data were collected in 2005 as a baseline of stream conditions for selected streams. Three streams were selected as representative of various stream types, and one drainage network, the Big Lake drainage basin, was selected for a systematic assessment. Streams in the Big Lake basin were drawn in a Geographic Information System (GIS), and 55 reaches along 16 miles of Meadow Creek and its primary tributary Little Meadow Creek were identified from orthoimagery and field observations on the basis of distinctive physical and habitat parameters, most commonly gradient, substrate, and vegetation. Data-collection methods for sites at the three representative reaches and the 55 systematically studied reaches consisted of a field survey of channel and flood-plain geometry and collection of 14 habitat attributes using published protocols or slight modifications. Width/depth and entrenchment ratios along the Meadow-Little Meadow Creek corridor were large and highly variable upstream of Parks Highway and lower and more consistent downstream of Parks Highway. Channel width was strongly correlated with distance, increasing downstream in a log-linear relation. Runs formed the most common habitat type, and instream vegetation dominated the habitat cover types, which collectively covered 53 percent of the channel. Gravel suitable for spawning covered isolated areas along Meadow Creek and about 29 percent of Little Meadow Creek. Broad wetlands were common along both streams. For a comprehensive assessment of small streams in the Mat-Su Valley, critical additional data needs include hydrologic, geologic and geomorphic, and biologic data

  1. The Pacific northwest stream quality assessment

    Science.gov (United States)

    Van Metre, Peter C.; Morace, Jennifer L.; Sheibley, Rich W.

    2015-01-01

    In 2015, the U.S. Geological Survey (USGS) National Water-Quality Assessment (NAWQA) program is assessing stream quality in the Pacific Northwest. The goals of the Pacific Northwest Stream Quality Assessment (Pacific Northwest study) are to assess the quality of streams in the region by characterizing multiple water-quality factors that are stressors to aquatic life and to evaluate the relation between these stressors and biological communities. The effects of urbanization and agriculture on stream quality for the Puget Lowlands and Willamette Valley are the focus of this regional study. Findings will provide the public and policymakers with information regarding which human and environmental factors are the most critical in affecting stream quality and, thus, provide insights about possible approaches to protect or improve the health of streams in the region.

  2. A New Study of Two Divergence Metrics for Change Detection in Data Streams

    KAUST Repository

    Qahtan, Abdulhakim Ali Ali; Wang, Suojin; Carroll, Raymond; Zhang, Xiangliang

    2014-01-01

    Streaming data are dynamic in nature with frequent changes. To detect such changes, most methods measure the difference between the data distributions in a current time window and a reference window. Divergence metrics and density estimation are required to measure the difference between the data distributions. Our study shows that the Kullback-Leibler (KL) divergence, the most popular metric for comparing distributions, fails to detect certain changes due to its asymmetric property and its dependence on the variance of the data. We thus consider two metrics for detecting changes in univariate data streams: a symmetric KL-divergence and a divergence metric measuring the intersection area of two distributions. The experimental results show that these two metrics lead to more accurate results in change detection than baseline methods such as Change Finder and using conventional KL-divergence.

  3. A New Study of Two Divergence Metrics for Change Detection in Data Streams

    KAUST Repository

    Qahtan, Abdulhakim Ali Ali

    2014-08-01

    Streaming data are dynamic in nature with frequent changes. To detect such changes, most methods measure the difference between the data distributions in a current time window and a reference window. Divergence metrics and density estimation are required to measure the difference between the data distributions. Our study shows that the Kullback-Leibler (KL) divergence, the most popular metric for comparing distributions, fails to detect certain changes due to its asymmetric property and its dependence on the variance of the data. We thus consider two metrics for detecting changes in univariate data streams: a symmetric KL-divergence and a divergence metric measuring the intersection area of two distributions. The experimental results show that these two metrics lead to more accurate results in change detection than baseline methods such as Change Finder and using conventional KL-divergence.

  4. High-speed packet filtering utilizing stream processors

    Science.gov (United States)

    Hummel, Richard J.; Fulp, Errin W.

    2009-04-01

    Parallel firewalls offer a scalable architecture for the next generation of high-speed networks. While these parallel systems can be implemented using multiple firewalls, the latest generation of stream processors can provide similar benefits with a significantly reduced latency due to locality. This paper describes how the Cell Broadband Engine (CBE), a popular stream processor, can be used as a high-speed packet filter. Results show the CBE can potentially process packets arriving at a rate of 1 Gbps with a latency less than 82 μ-seconds. Performance depends on how well the packet filtering process is translated to the unique stream processor architecture. For example the method used for transmitting data and control messages among the pseudo-independent processor cores has a significant impact on performance. Experimental results will also show the current limitations of a CBE operating system when used to process packets. Possible solutions to these issues will be discussed.

  5. Evaluating habitat associations of a fish assemblage at multiple spatial scales in a minimally disturbed stream using low-cost remote sensing

    Science.gov (United States)

    Cheek, Brandon D.; Grabowski, Timothy B.; Bean, Preston T.; Groeschel, Jillian R.; Magnelia, Stephan J.

    2016-01-01

    Habitat heterogeneity at multiple scales is a major factor affecting fish assemblage structure. However, assessments that examine these relationships at multiple scales concurrently are lacking. The lack of assessments at these scales is a critical gap in understanding as conservation and restoration efforts typically work at these levels.A combination of low-cost side-scan sonar surveys, aerial imagery using an unmanned aerial vehicle, and fish collections were used to evaluate the relationship between physicochemical and landscape variables at various spatial scales (e.g. micro-mesohabitat, mesohabitat, channel unit, stream reach) and stream–fish assemblage structure and habitat associations in the South Llano River, a spring-fed second-order stream on the Edwards Plateau in central Texas during 2012–2013.Low-cost side-scan sonar surveys have not typically been used to generate data for riverscape assessments of assemblage structure, thus the secondary objective was to assess the efficacy of this approach.The finest spatial scale (micro-mesohabitat) and the intermediate scale (channel unit) had the greatest explanatory power for variation in fish assemblage structure.Many of the fish endemic to the Edwards Plateau showed similar associations with physicochemical and landscape variables suggesting that conservation and restoration actions targeting a single endemic species may provide benefits to a large proportion of the endemic species in this system.Low-cost side-scan sonar proved to be a cost-effective means of acquiring information on the habitat availability of the entire river length and allowed the assessment of how a full suite of riverscape-level variables influenced local fish assemblage structure.

  6. Analysis of massive data streams using R and AMIDST

    DEFF Research Database (Denmark)

    Madsen, Anders Læsø; Salmerón, Antonio

    methods able to handle massive data streams based on Bayesian networks technology. All of the developed methods are available through the AMIDST toolbox, implemented in Java 8. We show how the functionality of the AMIDST toolbox can be accessed from R. Available AMIDST objects include variables......, distributions and Bayesian networks, as well as those devoted to inference and learning. The interaction between both platforms relies on the rJava package....

  7. Prioritized Contact Transport Stream

    Science.gov (United States)

    Hunt, Walter Lee, Jr. (Inventor)

    2015-01-01

    A detection process, contact recognition process, classification process, and identification process are applied to raw sensor data to produce an identified contact record set containing one or more identified contact records. A prioritization process is applied to the identified contact record set to assign a contact priority to each contact record in the identified contact record set. Data are removed from the contact records in the identified contact record set based on the contact priorities assigned to those contact records. A first contact stream is produced from the resulting contact records. The first contact stream is streamed in a contact transport stream. The contact transport stream may include and stream additional contact streams. The contact transport stream may be varied dynamically over time based on parameters such as available bandwidth, contact priority, presence/absence of contacts, system state, and configuration parameters.

  8. Flood-frequency characteristics of Wisconsin streams

    Science.gov (United States)

    Walker, John F.; Peppler, Marie C.; Danz, Mari E.; Hubbard, Laura E.

    2017-05-22

    Flood-frequency characteristics for 360 gaged sites on unregulated rural streams in Wisconsin are presented for percent annual exceedance probabilities ranging from 0.2 to 50 using a statewide skewness map developed for this report. Equations of the relations between flood-frequency and drainage-basin characteristics were developed by multiple-regression analyses. Flood-frequency characteristics for ungaged sites on unregulated, rural streams can be estimated by use of the equations presented in this report. The State was divided into eight areas of similar physiographic characteristics. The most significant basin characteristics are drainage area, soil saturated hydraulic conductivity, main-channel slope, and several land-use variables. The standard error of prediction for the equation for the 1-percent annual exceedance probability flood ranges from 56 to 70 percent for Wisconsin Streams; these values are larger than results presented in previous reports. The increase in the standard error of prediction is likely due to increased variability of the annual-peak discharges, resulting in increased variability in the magnitude of flood peaks at higher frequencies. For each of the unregulated rural streamflow-gaging stations, a weighted estimate based on the at-site log Pearson type III analysis and the multiple regression results was determined. The weighted estimate generally has a lower uncertainty than either the Log Pearson type III or multiple regression estimates. For regulated streams, a graphical method for estimating flood-frequency characteristics was developed from the relations of discharge and drainage area for selected annual exceedance probabilities. Graphs for the major regulated streams in Wisconsin are presented in the report.

  9. Multiple-jet thermal mixing in a piping tee

    International Nuclear Information System (INIS)

    Lykoudis, P.S.; Hagar, R.C.

    1979-01-01

    Piping tees that are used to mix fluid streams at different temperatures are subjected to possibly severe thermal and mechanical stresses. There is reason to suspect that mixing in a piping tee could be improved by injecting the fluid streams into the tee through multiple jets. This paper reports the results of an experimental investigation of the effects of multiple-jet injection on mixing in a piping tee. The experimental work involves the measurement of the temperature fluctuation intensity with a hot-film sensor downstream of a simple 22.22-mm(7/8-in.)-diam tee with mixed multiple-jet injected hot and cold streams of water. The jets were provided by holes drilled in plates that partially blocked the inlet streams; 26 pairs of plates were investigated. The number of holes per plate varied from 1 to 51; the jet diameters ranged from 5 to 68% of the tee diameter. The inlet stream Reynolds number upstream of the jet plates was roughly 15 500 for each stream. The data indicated that the root mean square (rms) temperature fluctuation intensity measured at the tee outlet decreased dramatically as the jet plate cross-sectional area void fraction was decreased. When the jets emanating from the tee plates were misaligned, the reduction of the rms temperature fluctuation was not as high as when the jets were aligned. The rate of decay of the intensity downstream of the tee for most ofthe plates investigated was found to agree well with the -3/4 power decay law predicted by Corrsin's theory of scalar decay. However, unusual features in the intensity decay data were also observed, such as an increase of the intensity several diameters downstream before continuing to decay

  10. Stream Classification Tool User Manual: For Use in Applications in Hydropower-Related Evironmental Mitigation

    Energy Technology Data Exchange (ETDEWEB)

    McManamay, Ryan A. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Troia, Matthew J. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); DeRolph, Christopher R. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Samu, Nicole M. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2016-01-01

    Stream classifications are an inventory of different types of streams. Classifications help us explore similarities and differences among different types of streams, make inferences regarding stream ecosystem behavior, and communicate the complexities of ecosystems. We developed a nested, layered, and spatially contiguous stream classification to characterize the biophysical settings of stream reaches within the Eastern United States (~ 900,000 reaches). The classification is composed of five natural characteristics (hydrology, temperature, size, confinement, and substrate) along with several disturbance regime layers, and each was selected because of their relevance to hydropower mitigation. We developed the classification at the stream reach level using the National Hydrography Dataset Plus Version 1 (1:100k scale). The stream classification is useful to environmental mitigation for hydropower dams in multiple ways. First, it creates efficiency in the regulatory process by creating an objective and data-rich means to address meaningful mitigation actions. Secondly, the SCT addresses data gaps as it quickly provides an inventory of hydrology, temperature, morphology, and ecological communities for the immediate project area, but also surrounding streams. This includes identifying potential reference streams as those that are proximate to the hydropower facility and fall within the same class. These streams can potentially be used to identify ideal environmental conditions or identify desired ecological communities. In doing so, the stream provides some context for how streams may function, respond to dam regulation, and an overview of specific mitigation needs. Herein, we describe the methodology in developing each stream classification layer and provide a tutorial to guide applications of the classification (and associated data) in regulatory settings, such as hydropower (re)licensing.

  11. Multicast Delayed Authentication For Streaming Synchrophasor Data in the Smart Grid.

    Science.gov (United States)

    Câmara, Sérgio; Anand, Dhananjay; Pillitteri, Victoria; Carmo, Luiz

    2016-01-01

    Multicast authentication of synchrophasor data is challenging due to the design requirements of Smart Grid monitoring systems such as low security overhead, tolerance of lossy networks, time-criticality and high data rates. In this work, we propose inf -TESLA, Infinite Timed Efficient Stream Loss-tolerant Authentication, a multicast delayed authentication protocol for communication links used to stream synchrophasor data for wide area control of electric power networks. Our approach is based on the authentication protocol TESLA but is augmented to accommodate high frequency transmissions of unbounded length. inf TESLA protocol utilizes the Dual Offset Key Chains mechanism to reduce authentication delay and computational cost associated with key chain commitment. We provide a description of the mechanism using two different modes for disclosing keys and demonstrate its security against a man-in-the-middle attack attempt. We compare our approach against the TESLA protocol in a 2-day simulation scenario, showing a reduction of 15.82% and 47.29% in computational cost, sender and receiver respectively, and a cumulative reduction in the communication overhead.

  12. Reducing False Negative Reads in RFID Data Streams Using an Adaptive Sliding-Window Approach

    Directory of Open Access Journals (Sweden)

    Herman Vermaak

    2012-03-01

    Full Text Available Unreliability of the data streams generated by RFID readers is among the primary factors which limit the widespread adoption of the RFID technology. RFID data cleaning is, therefore, an essential task in the RFID middleware systems in order to reduce reading errors, and to allow these data streams to be used to make a correct interpretation and analysis of the physical world they are representing. In this paper we propose an adaptive sliding-window based approach called WSTD which is capable of efficiently coping with both environmental variation and tag dynamics. Our experimental results demonstrate the efficacy of the proposed approach.

  13. Groundwater data improve modelling of headwater stream CO2 outgassing with a stable DIC isotope approach

    Science.gov (United States)

    Marx, Anne; Conrad, Marcus; Aizinger, Vadym; Prechtel, Alexander; van Geldern, Robert; Barth, Johannes A. C.

    2018-05-01

    A large portion of terrestrially derived carbon outgasses as carbon dioxide (CO2) from streams and rivers to the atmosphere. Particularly, the amount of CO2 outgassing from small headwater streams is highly uncertain. Conservative estimates suggest that they contribute 36 % (i.e. 0.93 petagrams (Pg) C yr-1) of total CO2 outgassing from all fluvial ecosystems on the globe. In this study, stream pCO2, dissolved inorganic carbon (DIC), and δ13CDIC data were used to determine CO2 outgassing from an acidic headwater stream in the Uhlířská catchment (Czech Republic). This stream drains a catchment with silicate bedrock. The applied stable isotope model is based on the principle that the 13C / 12C ratio of its sources and the intensity of CO2 outgassing control the isotope ratio of DIC in stream water. It avoids the use of the gas transfer velocity parameter (k), which is highly variable and mostly difficult to constrain. Model results indicate that CO2 outgassing contributed more than 80 % to the annual stream inorganic carbon loss in the Uhlířská catchment. This translated to a CO2 outgassing rate from the stream of 34.9 kg C m-2 yr-1 when normalised to the stream surface area. Large temporal variations with maximum values shortly before spring snowmelt and in summer emphasise the need for investigations at higher temporal resolution. We improved the model uncertainty by incorporating groundwater data to better constrain the isotope compositions of initial DIC. Due to the large global abundance of acidic, humic-rich headwaters, we underline the importance of this integral approach for global applications.

  14. Towards automatic parameter tuning of stream processing systems

    KAUST Repository

    Bilal, Muhammad

    2017-09-27

    Optimizing the performance of big-data streaming applications has become a daunting and time-consuming task: parameters may be tuned from a space of hundreds or even thousands of possible configurations. In this paper, we present a framework for automating parameter tuning for stream-processing systems. Our framework supports standard black-box optimization algorithms as well as a novel gray-box optimization algorithm. We demonstrate the multiple benefits of automated parameter tuning in optimizing three benchmark applications in Apache Storm. Our results show that a hill-climbing algorithm that uses a new heuristic sampling approach based on Latin Hypercube provides the best results. Our gray-box algorithm provides comparable results while being two to five times faster.

  15. LHCb trigger streams optimization

    Science.gov (United States)

    Derkach, D.; Kazeev, N.; Neychev, R.; Panin, A.; Trofimov, I.; Ustyuzhanin, A.; Vesterinen, M.

    2017-10-01

    The LHCb experiment stores around 1011 collision events per year. A typical physics analysis deals with a final sample of up to 107 events. Event preselection algorithms (lines) are used for data reduction. Since the data are stored in a format that requires sequential access, the lines are grouped into several output file streams, in order to increase the efficiency of user analysis jobs that read these data. The scheme efficiency heavily depends on the stream composition. By putting similar lines together and balancing the stream sizes it is possible to reduce the overhead. We present a method for finding an optimal stream composition. The method is applied to a part of the LHCb data (Turbo stream) on the stage where it is prepared for user physics analysis. This results in an expected improvement of 15% in the speed of user analysis jobs, and will be applied on data to be recorded in 2017.

  16. Development of a cross-section based stream package for MODFLOW

    Science.gov (United States)

    Ou, G.; Chen, X.; Irmak, A.

    2012-12-01

    Accurate simulation of stream-aquifer interactions for wide rivers using the streamflow routing package in MODFLOW is very challenging. To better represent a wide river spanning over multiple model grid cells, a Cross-Section based streamflow Routing (CSR) package is developed and incorporated into MODFLOW to simulate the interaction between streams and aquifers. In the CSR package, a stream segment is represented as a four-point polygon instead of a polyline which is traditionally used in streamflow routing simulation. Each stream segment is composed of upstream and downstream cross-sections. A cross-section consists of a number of streambed points possessing coordinates, streambed thicknesses and streambed hydraulic conductivities to describe the streambed geometry and hydraulic properties. The left and right end points are used to determine the locations of the stream segments. According to the cross-section geometry and hydraulic properties, CSR calculates the new stream stage at the cross-section using the Brent's method to solve the Manning's Equation. A module is developed to automatically compute the area of the stream segment polygon on each intersected MODFLOW grid cell as the upstream and downstream stages change. The stream stage and streambed hydraulic properties of model grids are interpolated based on the streambed points. Streambed leakage is computed as a function of streambed conductance and difference between the groundwater level and stream stage. The Muskingum-Cunge flow routing scheme with variable parameters is used to simulate the streamflow as the groundwater (discharge or recharge) contributes as lateral flows. An example is used to illustrate the capabilities of the CSR package. The result shows that the CSR is applicable to describing the spatial and temporal variation in the interaction between streams and aquifers. The input data become simple due to that the internal program automatically interpolates the cross-section data to each

  17. The Relationship Between Grazing, Er osion and Adult Aquatic Insects in Streams in Mongolia.

    Directory of Open Access Journals (Sweden)

    Barbara Hayford

    2010-06-01

    Full Text Available Overgrazing along stream channels in Mongolia may impact streams by increasing stream channel erosion and in-stream sediments, water temperature, pH, and conductivity. Grazing and erosion impacts may impair stream insects. The Mongolian Aquatic Insect Survey sampled 250 streams during summer seasons in 2003-2006 and 2008. On-site identifi cations of aquatic insect families mostly based on collections of adults were recorded for each site, leading us to ask whether the family-level data were useful in biological assessment related to impacts and impairment from grazing and erosion. A double dendrogram based on hierarchical cluster analysis was used to fi nd patterns in sites and aquatic insect communities. Sites did not group by sampling period, but some sites did group by stream size and elevation. However, elevation was not a signifi cant predictor of variation in aquatic insect metrics. Analysis of variance was used to determine whether insect metrics and water quality variables varied signifi cantly between categories of erosion in the stream channel. Plecoptera and Diptera richness decreased with increased erosion and Percent Diptera Richness was the only aquatic insect metric to vary signifi cantly between categories of erosion along the stream channel. Water temperature, conductivity, and pH also signifi cantly increased with increased erosion. Multiple regression analysis was used to determine whether aquatic insect metrics could be predicted by variation in landscape, water quality and stream reach variables. Trichoptera, Ephemeroptera, and Coleoptera richness increased with increased erosion, conductivity, and pH, but not signifi cantly. Percent Diptera Richness formed the only signifi cant model in the multiple regression analysis, with conductivity the only signifi cant predictor of variation in Percent Diptera Richness. Family-level data generated in the fi eld indicated that sampling for Trichoptera and Ephemeroptera diversity would

  18. Groundwater data improve modelling of headwater stream CO2 outgassing with a stable DIC isotope approach

    Directory of Open Access Journals (Sweden)

    A. Marx

    2018-05-01

    Full Text Available A large portion of terrestrially derived carbon outgasses as carbon dioxide (CO2 from streams and rivers to the atmosphere. Particularly, the amount of CO2 outgassing from small headwater streams is highly uncertain. Conservative estimates suggest that they contribute 36 % (i.e. 0.93 petagrams (Pg C yr−1 of total CO2 outgassing from all fluvial ecosystems on the globe. In this study, stream pCO2, dissolved inorganic carbon (DIC, and δ13CDIC data were used to determine CO2 outgassing from an acidic headwater stream in the Uhlířská catchment (Czech Republic. This stream drains a catchment with silicate bedrock. The applied stable isotope model is based on the principle that the 13C ∕ 12C ratio of its sources and the intensity of CO2 outgassing control the isotope ratio of DIC in stream water. It avoids the use of the gas transfer velocity parameter (k, which is highly variable and mostly difficult to constrain. Model results indicate that CO2 outgassing contributed more than 80 % to the annual stream inorganic carbon loss in the Uhlířská catchment. This translated to a CO2 outgassing rate from the stream of 34.9 kg C m−2 yr−1 when normalised to the stream surface area. Large temporal variations with maximum values shortly before spring snowmelt and in summer emphasise the need for investigations at higher temporal resolution. We improved the model uncertainty by incorporating groundwater data to better constrain the isotope compositions of initial DIC. Due to the large global abundance of acidic, humic-rich headwaters, we underline the importance of this integral approach for global applications.

  19. Ensemble Classification of Data Streams Based on Attribute Reduction and a Sliding Window

    Directory of Open Access Journals (Sweden)

    Yingchun Chen

    2018-04-01

    Full Text Available With the current increasing volume and dimensionality of data, traditional data classification algorithms are unable to satisfy the demands of practical classification applications of data streams. To deal with noise and concept drift in data streams, we propose an ensemble classification algorithm based on attribute reduction and a sliding window in this paper. Using mutual information, an approximate attribute reduction algorithm based on rough sets is used to reduce data dimensionality and increase the diversity of reduced results in the algorithm. A double-threshold concept drift detection method and a three-stage sliding window control strategy are introduced to improve the performance of the algorithm when dealing with both noise and concept drift. The classification precision is further improved by updating the base classifiers and their nonlinear weights. Experiments on synthetic datasets and actual datasets demonstrate the performance of the algorithm in terms of classification precision, memory use, and time efficiency.

  20. Using macroinvertebrate assemblages and multiple stressors to infer urban stream system condition: A case study in the central US

    Science.gov (United States)

    Nichols, John W.; Hubbart, Jason A.; Poulton, Barry C.

    2016-01-01

    Characterizing the impacts of hydrologic alterations, pollutants, and habitat degradation on macroinvertebrate species assemblages is of critical value for managers wishing to categorize stream ecosystem condition. A combination of approaches including trait-based metrics and traditional bioassessments provides greater information, particularly in anthropogenic stream ecosystems where traditional approaches can be confounded by variously interacting land use impacts. Macroinvertebrates were collected from two rural and three urban nested study sites in central Missouri, USA during the spring and fall seasons of 2011. Land use responses of conventional taxonomic and trait-based metrics were compared to streamflow indices, physical habitat metrics, and water quality indices. Results show that biotic index was significantly different (p habitats in urban reaches contained 21 % more (p = 0.03) multivoltine organisms, which was positively correlated to the magnitude of peak flows (r2 = 0.91, p = 0.012) suggesting that high flow events may serve as a disturbance in those areas. Results support the use of macroinvertebrate assemblages and multiple stressors to characterize urban stream system condition and highlight the need to better understand the complex interactions of trait-based metrics and anthropogenic aquatic ecosystem stressors.

  1. The Northeast Stream Quality Assessment

    Science.gov (United States)

    Van Metre, Peter C.; Riva-Murray, Karen; Coles, James F.

    2016-04-22

    In 2016, the U.S. Geological Survey (USGS) National Water-Quality Assessment (NAWQA) is assessing stream quality in the northeastern United States. The goal of the Northeast Stream Quality Assessment (NESQA) is to assess the quality of streams in the region by characterizing multiple water-quality factors that are stressors to aquatic life and evaluating the relation between these stressors and biological communities. The focus of NESQA in 2016 will be on the effects of urbanization and agriculture on stream quality in all or parts of eight states: Connecticut, Massachusetts, New Hampshire, New Jersey, New York, Pennsylvania, Rhode Island, and Vermont.Findings will provide the public and policymakers with information about the most critical factors affecting stream quality, thus providing insights about possible approaches to protect the health of streams in the region. The NESQA study will be the fourth regional study conducted as part of NAWQA and will be of similar design and scope to the first three, in the Midwest in 2013, the Southeast in 2014, and the Pacific Northwest in 2015 (http://txpub.usgs.gov/RSQA/).

  2. Hydrogeologic data for the Big River-Mishnock River stream-aquifer system, central Rhode Island

    Science.gov (United States)

    Craft, P.A.

    2001-01-01

    Hydrogeology, ground-water development alternatives, and water quality in the BigMishnock stream-aquifer system in central Rhode Island are being investigated as part of a long-term cooperative program between the Rhode Island Water Resources Board and the U.S. Geological Survey to evaluate the ground-water resources throughout Rhode Island. The study area includes the Big River drainage basin and that portion of the Mishnock River drainage basin upstream from the Mishnock River at State Route 3. This report presents geologic data and hydrologic and water-quality data for ground and surface water. Ground-water data were collected from July 1996 through September 1998 from a network of observation wells consisting of existing wells and wells installed for this study, which provided a broad distribution of data-collection sites throughout the study area. Streambed piezometers were used to obtain differences in head data between surface-water levels and ground-water levels to help evaluate stream-aquifer interactions throughout the study area. The types of data presented include monthly ground-water levels, average daily ground-water withdrawals, drawdown data from aquifer tests, and water-quality data. Historical water-level data from other wells within the study area also are presented in this report. Surface-water data were obtained from a network consisting of surface-water impoundments, such as ponds and reservoirs, existing and newly established partial-record stream-discharge sites, and synoptic surface-water-quality sites. Water levels were collected monthly from the surface-water impoundments. Stream-discharge measurements were made at partial-record sites to provide measurements of inflow, outflow, and internal flow throughout the study area. Specific conductance was measured monthly at partial-record sites during the study, and also during the fall and spring of 1997 and 1998 at 41 synoptic sites throughout the study area. General geologic data, such as

  3. Low-dimensional and Data Fusion Techniques Applied to a Rectangular Supersonic Multi-stream Jet

    Science.gov (United States)

    Berry, Matthew; Stack, Cory; Magstadt, Andrew; Ali, Mohd; Gaitonde, Datta; Glauser, Mark

    2017-11-01

    Low-dimensional models of experimental and simulation data for a complex supersonic jet were fused to reconstruct time-dependent proper orthogonal decomposition (POD) coefficients. The jet consists of a multi-stream rectangular single expansion ramp nozzle, containing a core stream operating at Mj , 1 = 1.6 , and bypass stream at Mj , 3 = 1.0 with an underlying deck. POD was applied to schlieren and PIV data to acquire the spatial basis functions. These eigenfunctions were projected onto their corresponding time-dependent large eddy simulation (LES) fields to reconstruct the temporal POD coefficients. This reconstruction was able to resolve spectral peaks that were previously aliased due to the slower sampling rates of the experiments. Additionally, dynamic mode decomposition (DMD) was applied to the experimental and LES datasets, and the spatio-temporal characteristics were compared to POD. The authors would like to acknowledge AFOSR, program manager Dr. Doug Smith, for funding this research, Grant No. FA9550-15-1-0435.

  4. Adaptive Framework for Classification and Novel Class Detection over Evolving Data Streams with Limited Labeled Data.

    Energy Technology Data Exchange (ETDEWEB)

    Haque, Ahsanul [Univ. of Texas, Dallas, TX (United States); Khan, Latifur [Univ. of Texas, Dallas, TX (United States); Baron, Michael [Univ. of Texas, Dallas, TX (United States); Ingram, Joey Burton [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-09-01

    Most approaches to classifying evolving data streams either divide the stream of data into fixed-size chunks or use gradual forgetting to address the problems of infinite length and concept drift. Finding the fixed size of the chunks or choosing a forgetting rate without prior knowledge about time-scale of change is not a trivial task. As a result, these approaches suffer from a trade-off between performance and sensitivity. To address this problem, we present a framework which uses change detection techniques on the classifier performance to determine chunk boundaries dynamically. Though this framework exhibits good performance, it is heavily dependent on the availability of true labels of data instances. However, labeled data instances are scarce in realistic settings and not readily available. Therefore, we present a second framework which is unsupervised in nature, and exploits change detection on classifier confidence values to determine chunk boundaries dynamically. In this way, it avoids the use of labeled data while still addressing the problems of infinite length and concept drift. Moreover, both of our proposed frameworks address the concept evolution problem by detecting outliers having similar values for the attributes. We provide theoretical proof that our change detection method works better than other state-of-the-art approaches in this particular scenario. Results from experiments on various benchmark and synthetic data sets also show the efficiency of our proposed frameworks.

  5. Analysis of hydraulic characteristics for stream diversion in small stream

    Energy Technology Data Exchange (ETDEWEB)

    Ahn, Sang-Jin; Jun, Kye-Won [Chungbuk National University, Cheongju(Korea)

    2001-10-31

    This study is the analysis of hydraulic characteristics for stream diversion reach by numerical model test. Through it we can provide the basis data in flood, and in grasping stream flow characteristics. Analysis of hydraulic characteristics in Seoknam stream were implemented by using computer model HEC-RAS(one-dimensional model) and RMA2(two-dimensional finite element model). As a result we became to know that RMA2 to simulate left, main channel, right in stream is more effective method in analysing flow in channel bends, steep slope, complex bed form effect stream flow characteristics, than HEC-RAS. (author). 13 refs., 3 tabs., 5 figs.

  6. Short-term stream flow forecasting at Australian river sites using data-driven regression techniques

    CSIR Research Space (South Africa)

    Steyn, Melise

    2017-09-01

    Full Text Available This study proposes a computationally efficient solution to stream flow forecasting for river basins where historical time series data are available. Two data-driven modeling techniques are investigated, namely support vector regression...

  7. The end of polling: why and how to transform a REST API into a Data Streaming API?

    CERN Multimedia

    CERN. Geneva

    2017-01-01

    We know interactivity is the key to keep our user’s interest alive but we can’t reduce animation to UI anymore. Twitter, Waze, Slack… users are used to have real-time data in applications they love. But how can you turn your static API into a stream of data? By pulling? Pushing? Webhook-ing? When talking about data streaming, we often think about WebSockets. But have you ever heard of Server-Sent Events? In this tools-in-action we will compare those technologies to understand which one you should opt for depending on your usecase, and I’ll show you how we have been reducing the amount of data to transfer even further with JSON-Patch. And because real-time data is not only needed by web (and because it’s much more fun), I’ll show you how we can make a drone dance on streamed APIs.

  8. Deep SOMs for automated feature extraction and classification from big data streaming

    Science.gov (United States)

    Sakkari, Mohamed; Ejbali, Ridha; Zaied, Mourad

    2017-03-01

    In this paper, we proposed a deep self-organizing map model (Deep-SOMs) for automated features extracting and learning from big data streaming which we benefit from the framework Spark for real time streams and highly parallel data processing. The SOMs deep architecture is based on the notion of abstraction (patterns automatically extract from the raw data, from the less to more abstract). The proposed model consists of three hidden self-organizing layers, an input and an output layer. Each layer is made up of a multitude of SOMs, each map only focusing at local headmistress sub-region from the input image. Then, each layer trains the local information to generate more overall information in the higher layer. The proposed Deep-SOMs model is unique in terms of the layers architecture, the SOMs sampling method and learning. During the learning stage we use a set of unsupervised SOMs for feature extraction. We validate the effectiveness of our approach on large data sets such as Leukemia dataset and SRBCT. Results of comparison have shown that the Deep-SOMs model performs better than many existing algorithms for images classification.

  9. A stream temperature model for the Peace-Athabasca River basin

    Science.gov (United States)

    Morales-Marin, L. A.; Rokaya, P.; Wheater, H. S.; Lindenschmidt, K. E.

    2017-12-01

    Water temperature plays a fundamental role in water ecosystem functioning. Because it regulates flow energy and metabolic rates in organism productivity over a broad spectrum of space and time scales, water temperature constitutes an important indicator of aquatic ecosystems health. In cold region basins, stream water temperature modelling is also fundamental to predict ice freeze-up and break-up events in order to improve flood management. Multiple model approaches such as linear and multivariable regression methods, neural network and thermal energy budged models have been developed and implemented to simulate stream water temperature. Most of these models have been applied to specific stream reaches and trained using observed data, but very little has been done to simulate water temperature in large catchment river networks. We present the coupling of RBM model, a semi-Lagrangian water temperature model for advection-dominated river system, and MESH, a semi-distributed hydrological model, to simulate stream water temperature in river catchments. The coupled models are implemented in the Peace-Athabasca River basin in order to analyze the variation in stream temperature regimes under changing hydrological and meteorological conditions. Uncertainty of stream temperature simulations is also assessed in order to determine the degree of reliability of the estimates.

  10. Energy Efficiency Effects of Vectorization in Data Reuse Transformations for Many-Core Processors—A Case Study †

    Directory of Open Access Journals (Sweden)

    Abdullah Al Hasib

    2017-02-01

    Full Text Available Thread-level and data-level parallel architectures have become the design of choice in many of today’s energy-efficient computing systems. However, these architectures put substantially higher requirements on the memory subsystem than scalar architectures, making memory latency and bandwidth critical in their overall efficiency. Data reuse exploration aims at reducing the pressure on the memory subsystem by exploiting the temporal locality in data accesses. In this paper, we investigate the effects on performance and energy from a data reuse methodology combined with parallelization and vectorization in multi- and many-core processors. As a test case, a full-search motion estimation kernel is evaluated on Intel® CoreTM i7-4700K (Haswell and i7-2600K (Sandy Bridge multi-core processors, as well as on an Intel® Xeon PhiTM many-core processor (Knights Landing with Streaming Single Instruction Multiple Data (SIMD Extensions (SSE and Advanced Vector Extensions (AVX instruction sets. Results using a single-threaded execution on the Haswell and Sandy Bridge systems show that performance and EDP (Energy Delay Product can be improved through data reuse transformations on the scalar code by a factor of ≈3× and ≈6×, respectively. Compared to scalar code without data reuse optimization, the SSE/AVX2 version achieves ≈10×/17× better performance and ≈92×/307× better EDP, respectively. These results can be improved by 10% to 15% using data reuse techniques. Finally, the most optimized version using data reuse and AVX512 achieves a speedup of ≈35× and an EDP improvement of ≈1192× on the Xeon Phi system. While single-threaded execution serves as a common reference point for all architectures to analyze the effects of data reuse on both scalar and vector codes, scalability with thread count is also discussed in the paper.

  11. Sensitivity Sampling Over Dynamic Geometric Data Streams with Applications to $k$-Clustering

    OpenAIRE

    Song, Zhao; Yang, Lin F.; Zhong, Peilin

    2018-01-01

    Sensitivity based sampling is crucial for constructing nearly-optimal coreset for $k$-means / median clustering. In this paper, we provide a novel data structure that enables sensitivity sampling over a dynamic data stream, where points from a high dimensional discrete Euclidean space can be either inserted or deleted. Based on this data structure, we provide a one-pass coreset construction for $k$-means %and M-estimator clustering using space $\\widetilde{O}(k\\mathrm{poly}(d))$ over $d$-dimen...

  12. Data Reduction of Laser Ablation Split-Stream (LASS) Analyses Using Newly Developed Features Within Iolite: With Applications to Lu-Hf + U-Pb in Detrital Zircon and Sm-Nd +U-Pb in Igneous Monazite

    Science.gov (United States)

    Fisher, Christopher M.; Paton, Chad; Pearson, D. Graham; Sarkar, Chiranjeeb; Luo, Yan; Tersmette, Daniel B.; Chacko, Thomas

    2017-12-01

    A robust platform to view and integrate multiple data sets collected simultaneously is required to realize the utility and potential of the Laser Ablation Split-Stream (LASS) method. This capability, until now, has been unavailable and practitioners have had to laboriously process each data set separately, making it challenging to take full advantage of the benefits of LASS. We describe a new program for handling multiple mass spectrometric data sets collected simultaneously, designed specifically for the LASS technique, by which a laser aerosol is been split into two or more separate "streams" to be measured on separate mass spectrometers. New features within Iolite (https://iolite-software.com) enable the capability of loading, synchronizing, viewing, and reducing two or more data sets acquired simultaneously, as multiple DRSs (data reduction schemes) can be run concurrently. While this version of Iolite accommodates any combination of simultaneously collected mass spectrometer data, we demonstrate the utility using case studies where U-Pb and Lu-Hf isotope composition of zircon, and U-Pb and Sm-Nd isotope composition of monazite were analyzed simultaneously, in crystals showing complex isotopic zonation. These studies demonstrate the importance of being able to view and integrate simultaneously acquired data sets, especially for samples with complicated zoning and decoupled isotope systematics, in order to extract accurate and geologically meaningful isotopic and compositional data. This contribution provides instructions and examples for handling simultaneously collected laser ablation data. An instructional video is also provided. The updated Iolite software will help to fully develop the applications of both LASS and multi-instrument mass spectrometric measurement capabilities.

  13. Methods of natural gas liquefaction and natural gas liquefaction plants utilizing multiple and varying gas streams

    Science.gov (United States)

    Wilding, Bruce M; Turner, Terry D

    2014-12-02

    A method of natural gas liquefaction may include cooling a gaseous NG process stream to form a liquid NG process stream. The method may further include directing the first tail gas stream out of a plant at a first pressure and directing a second tail gas stream out of the plant at a second pressure. An additional method of natural gas liquefaction may include separating CO.sub.2 from a liquid NG process stream and processing the CO.sub.2 to provide a CO.sub.2 product stream. Another method of natural gas liquefaction may include combining a marginal gaseous NG process stream with a secondary substantially pure NG stream to provide an improved gaseous NG process stream. Additionally, a NG liquefaction plant may include a first tail gas outlet, and at least a second tail gas outlet, the at least a second tail gas outlet separate from the first tail gas outlet.

  14. Models of Tidally Induced Gas Filaments in the Magellanic Stream

    Science.gov (United States)

    Pardy, Stephen A.; D’Onghia, Elena; Fox, Andrew J.

    2018-04-01

    The Magellanic Stream and Leading Arm of H I that stretches from the Large and Small Magellanic Clouds (LMC and SMC) and over 200° of the Southern sky is thought to be formed from multiple encounters between the LMC and SMC. In this scenario, most of the gas in the Stream and Leading Arm is stripped from the SMC, yet recent observations have shown a bifurcation of the Trailing Arm that reveals LMC origins for some of the gas. Absorption measurements in the Stream also reveal an order of magnitude more gas than in current tidal models. We present hydrodynamical simulations of the multiple encounters between the LMC and SMC at their first pass around the Milky Way, assuming that the Clouds were more extended and gas-rich in the past. Our models create filamentary structures of gas in the Trailing Stream from both the LMC and SMC. While the SMC trailing filament matches the observed Stream location, the LMC filament is offset. In addition, the total observed mass of the Stream in these models is underestimated by a factor of four when the ionized component is accounted for. Our results suggest that there should also be gas stripped from both the LMC and SMC in the Leading Arm, mirroring the bifurcation in the Trailing Stream. This prediction is consistent with recent measurements of spatial variation in chemical abundances in the Leading Arm, which show that gas from multiple sources is present, although its nature is still uncertain.

  15. Linking land-use type and stream water quality using spatial data of fecal indicator bacteria and heavy metals in the Yeongsan river basin.

    Science.gov (United States)

    Kang, Joo-Hyon; Lee, Seung Won; Cho, Kyung Hwa; Ki, Seo Jin; Cha, Sung Min; Kim, Joon Ha

    2010-07-01

    This study reveals land-use factors that explain stream water quality during wet and dry weather conditions in a large river basin using two different linear models-multiple linear regression (MLR) models and constrained least squares (CLS) models. Six land-use types and three topographical parameters (size, slope, and permeability) of the watershed were incorporated into the models as explanatory variables. The suggested models were then demonstrated using a digitized elevation map in conjunction with the land-use and the measured concentration data for Escherichia coli (EC), Enterococci bacteria (ENT), and six heavy metal species collected monthly during 2007-2008 at 50 monitoring sites in the Yeongsan Watershed, Korea. The results showed that the MLR models can be a powerful tool for predicting the average concentrations of pollutants in stream water (the Nash-Sutcliffe (NS) model efficiency coefficients ranged from 0.67 to 0.95). On the other hand, the CLS models, with moderately good prediction performance (the NS coefficients ranged 0.28-0.85), were more suitable for quantifying contributions of respective land-uses to the stream water quality. The CLS models suggested that industrial and urban land-uses are major contributors to the stream concentrations of EC and ENT, whereas agricultural, industrial, and mining areas were significant sources of many heavy metal species. In addition, the slope, size, and permeability of the watershed were found to be important factors determining the extent of the contribution from each land-use type to the stream water quality. The models proposed in this paper can be considered useful tools for developing land cover guidelines and for prioritizing locations for implementing management practices to maintain stream water quality standard in a large river basin. Copyright 2010 Elsevier Ltd. All rights reserved.

  16. The role of international policy transfer within the Multiple Streams Approach: the case of smart electricity metering in Australia

    OpenAIRE

    Lovell, Heather

    2016-01-01

    This paper draws on Kingdon’s Multiple Streams Approach (MSA) to consider international flows of policy, not just domestic. It is argued that using the MSA in conjunction with international policy transfer and mobility theories allows for a fuller explanation of the development of smart electricity metering policy in Australia. The MSA is based originally on empirical research within a single country - the USA - in the late 1970s, and all three of the ‘streams’ identified as important to poli...

  17. Spatial patterns of stream temperatures and electric conductivity in a mesoscale catchment

    Science.gov (United States)

    Lieder, Ernestine; Weiler, Markus; Blume, Theresa

    2017-04-01

    Stream temperature and electric conductivity (EC) are both relatively easily measured and can provide valuable information on runoff generation processes and catchment storage.This study investigates the spatial variability of stream temperature and EC in a mesoscale basin. We focus on the mesoscale (sub-catchments and reach scale), and long term (seasonal / annual) stream temperature and EC patterns. Our study basin is the Attert catchment in Luxembourg (288km2), which contains multiple sub-catchments of different geology, topography and land use patterns. We installed 90 stream temperature and EC sensors at sites across the basin in summer 2015. The collected data is complemented by land use and discharge data and an extensive climate data set. Thermal sensitivity was calculated as the slope of daily air temperature-water-temperature regression line and describes the sensitivity of stream temperature to long term environmental change. Amplitude sensitivity was calculated as slope of the daily air and water temperature amplitude regression and describes the short term warming capacity of the stream. We found that groups with similar long term thermal and EC patterns are strongly related to different geological units. The sandstone reaches show the coldest temperatures and lowest annual thermal sensitivity to air temperature. The slate reaches are characterized by comparably low EC and high daily temperature amplitudes and amplitude sensitivity. Furthermore, mean annual temperatures and thermal sensitivities increase exponentially with drainage area, which can be attributed to the accumulation of heat throughout the system. On the reach scale, daily stream temperature fluctuations or sensitivities were strongly influenced by land cover distribution, stream shading and runoff volume. Daily thermal sensitivities were low for headwater streams; peaked for intermediate reaches in the middle of the catchment and then decreased again further downstream with increasing

  18. Jet stream winds - Comparisons of analyses with independent aircraft data over Southwest Asia

    Science.gov (United States)

    Tenenbaum, J.

    1991-01-01

    Cruise-level wind data from commercial aircraft are obtained, and these data are compared with operational jet stream analyses over southwest Asia, an area of limited conventional data. Results from an ensemble of 11 cases during January 1989 and individual cases during December 1988-March 1989 are presented. The key results are: (1) European Centre for Medium-Range Weather Forecasts (ECMWF), National Meteorological Center, and United Kingdom Meteorological Office analyses of the subtropical jet in southwest Asia are 11 percent, 17 percent, and 17 percent weaker, respectively, than aircraft observations; (2) analyzed poleward shears range up to 1 f (0.00007/s) compared with up to 3 f (0.00021/s) in the aircraft observations where f is the local Coriolis parameters; (3) the ECMWF errors are largest at the base of the jet; (4) the mean ECMWF core location is latitudinally correct but has an rms latitude variance of 1.5 deg; (5) isolated erroneous radiosondes produce unmeteorological structures in the analyzed subtropical jet stream; and (6) the increased utilization of automated aircraft reports is likely to produce a spurious secular increase in the apparent strength of the jets. The magnitude and spatial extent of the errors seen are near limits of current GCM resolution (100 km) but should be resolvable. The results imply that studies of GCM systematic jet stream wind errors in weather and climate forecasts must be interpreted with caution in this region.

  19. Reconfigurable Multicore Architectures for Streaming Applications

    NARCIS (Netherlands)

    Smit, Gerardus Johannes Maria; Kokkeler, Andre B.J.; Rauwerda, G.K.; Jacobs, J.W.M.; Nicolescu, G.; Mosterman, P.J.

    2009-01-01

    This chapter addresses reconfigurable heterogenous and homogeneous multicore system-on-chip (SoC) platforms for streaming digital signal processing applications, also called DSP applications. In streaming DSP applications, computations can be specified as a data flow graph with streams of data items

  20. Data Centric Sensor Stream Reduction for Real-Time Applications in Wireless Sensor Networks

    Science.gov (United States)

    Aquino, Andre Luiz Lins; Nakamura, Eduardo Freire

    2009-01-01

    This work presents a data-centric strategy to meet deadlines in soft real-time applications in wireless sensor networks. This strategy considers three main aspects: (i) The design of real-time application to obtain the minimum deadlines; (ii) An analytic model to estimate the ideal sample size used by data-reduction algorithms; and (iii) Two data-centric stream-based sampling algorithms to perform data reduction whenever necessary. Simulation results show that our data-centric strategies meet deadlines without loosing data representativeness. PMID:22303145

  1. Unequal Protection of Video Streaming through Adaptive Modulation with a Trizone Buffer over Bluetooth Enhanced Data Rate

    Directory of Open Access Journals (Sweden)

    Razavi Rouzbeh

    2008-01-01

    Full Text Available Abstract Bluetooth enhanced data rate wireless channel can support higher-quality video streams compared to previous versions of Bluetooth. Packet loss when transmitting compressed data has an effect on the delivered video quality that endures over multiple frames. To reduce the impact of radio frequency noise and interference, this paper proposes adaptive modulation based on content type at the video frame level and content importance at the macroblock level. Because the bit rate of protected data is reduced, the paper proposes buffer management to reduce the risk of buffer overflow. A trizone buffer is introduced, with a varying unequal protection policy in each zone. Application of this policy together with adaptive modulation results in up to 4 dB improvement in objective video quality compared to fixed rate scheme for an additive white Gaussian noise channel and around 10 dB for a Gilbert-Elliott channel. The paper also reports a consistent improvement in video quality over a scheme that adapts to channel conditions by varying the data rate without accounting for the video frame packet type or buffer congestion.

  2. Unequal Protection of Video Streaming through Adaptive Modulation with a Trizone Buffer over Bluetooth Enhanced Data Rate

    Directory of Open Access Journals (Sweden)

    Rouzbeh Razavi

    2007-12-01

    Full Text Available Bluetooth enhanced data rate wireless channel can support higher-quality video streams compared to previous versions of Bluetooth. Packet loss when transmitting compressed data has an effect on the delivered video quality that endures over multiple frames. To reduce the impact of radio frequency noise and interference, this paper proposes adaptive modulation based on content type at the video frame level and content importance at the macroblock level. Because the bit rate of protected data is reduced, the paper proposes buffer management to reduce the risk of buffer overflow. A trizone buffer is introduced, with a varying unequal protection policy in each zone. Application of this policy together with adaptive modulation results in up to 4 dB improvement in objective video quality compared to fixed rate scheme for an additive white Gaussian noise channel and around 10 dB for a Gilbert-Elliott channel. The paper also reports a consistent improvement in video quality over a scheme that adapts to channel conditions by varying the data rate without accounting for the video frame packet type or buffer congestion.

  3. Validating alternative methodologies to estimate the hydrological regime of temporary streams when flow data are unavailable

    Science.gov (United States)

    Llorens, Pilar; Gallart, Francesc; Latron, Jérôme; Cid, Núria; Rieradevall, Maria; Prat, Narcís

    2016-04-01

    Aquatic life in temporary streams is strongly conditioned by the temporal variability of the hydrological conditions that control the occurrence and connectivity of diverse mesohabitats. In this context, the software TREHS (Temporary Rivers' Ecological and Hydrological Status) has been developed, in the framework of the LIFE Trivers project, to help managers for adequately implement the Water Framework Directive in this type of water bodies. TREHS, using the methodology described in Gallart et al (2012), defines six temporal 'aquatic states', based on the hydrological conditions representing different mesohabitats, for a given reach at a particular moment. Nevertheless, hydrological data for assessing the regime of temporary streams are often non-existent or scarce. The scarcity of flow data makes frequently impossible the characterization of temporary streams hydrological regimes and, as a consequence, the selection of the correct periods and methods to determine their ecological status. Because of its qualitative nature, the TREHS approach allows the use of alternative methodologies to assess the regime of temporary streams in the lack of observed flow data. However, to adapt the TREHS to this qualitative data both the temporal scheme (from monthly to seasonal) as well as the number of aquatic states (from 6 to 3) have been modified. Two alternatives complementary methodologies were tested within the TREHS framework to assess the regime of temporary streams: interviews and aerial photographs. All the gauging stations (13) belonging to the Catalan Internal Catchments (NE, Spain) with recurrent zero flows periods were selected to validate both methodologies. On one hand, non-structured interviews were carried out to inhabitants of villages and small towns near the gauging stations. Flow permanence metrics for input into TREHS were drawn from the notes taken during the interviews. On the other hand, the historical series of available aerial photographs (typically 10

  4. Grid-based Parallel Data Streaming Implemented for the Gyrokinetic Toroidal Code

    International Nuclear Information System (INIS)

    Klasky, S.; Ethier, S.; Lin, Z.; Martins, K.; McCune, D.; Samtaney, R.

    2003-01-01

    We have developed a threaded parallel data streaming approach using Globus to transfer multi-terabyte simulation data from a remote supercomputer to the scientist's home analysis/visualization cluster, as the simulation executes, with negligible overhead. Data transfer experiments show that this concurrent data transfer approach is more favorable compared with writing to local disk and then transferring this data to be post-processed. The present approach is conducive to using the grid to pipeline the simulation with post-processing and visualization. We have applied this method to the Gyrokinetic Toroidal Code (GTC), a 3-dimensional particle-in-cell code used to study microturbulence in magnetic confinement fusion from first principles plasma theory

  5. Parasail: SIMD C library for global, semi-global, and local pairwise sequence alignments.

    Science.gov (United States)

    Daily, Jeff

    2016-02-10

    Sequence alignment algorithms are a key component of many bioinformatics applications. Though various fast Smith-Waterman local sequence alignment implementations have been developed for x86 CPUs, most are embedded into larger database search tools. In addition, fast implementations of Needleman-Wunsch global sequence alignment and its semi-global variants are not as widespread. This article presents the first software library for local, global, and semi-global pairwise intra-sequence alignments and improves the performance of previous intra-sequence implementations. A faster intra-sequence local pairwise alignment implementation is described and benchmarked, including new global and semi-global variants. Using a 375 residue query sequence a speed of 136 billion cell updates per second (GCUPS) was achieved on a dual Intel Xeon E5-2670 24-core processor system, the highest reported for an implementation based on Farrar's 'striped' approach. Rognes's SWIPE optimal database search application is still generally the fastest available at 1.2 to at best 2.4 times faster than Parasail for sequences shorter than 500 amino acids. However, Parasail was faster for longer sequences. For global alignments, Parasail's prefix scan implementation is generally the fastest, faster even than Farrar's 'striped' approach, however the opal library is faster for single-threaded applications. The software library is designed for 64 bit Linux, OS X, or Windows on processors with SSE2, SSE41, or AVX2. Source code is available from https://github.com/jeffdaily/parasail under the Battelle BSD-style license. Applications that require optimal alignment scores could benefit from the improved performance. For the first time, SIMD global, semi-global, and local alignments are available in a stand-alone C library.

  6. Matisse: A Visual Analytics System for Exploring Emotion Trends in Social Media Text Streams

    Energy Technology Data Exchange (ETDEWEB)

    Steed, Chad A [ORNL; Drouhard, Margaret MEG G [ORNL; Beaver, Justin M [ORNL; Pyle, Joshua M [ORNL; BogenII, Paul L. [Google Inc.

    2015-01-01

    Dynamically mining textual information streams to gain real-time situational awareness is especially challenging with social media systems where throughput and velocity properties push the limits of a static analytical approach. In this paper, we describe an interactive visual analytics system, called Matisse, that aids with the discovery and investigation of trends in streaming text. Matisse addresses the challenges inherent to text stream mining through the following technical contributions: (1) robust stream data management, (2) automated sentiment/emotion analytics, (3) interactive coordinated visualizations, and (4) a flexible drill-down interaction scheme that accesses multiple levels of detail. In addition to positive/negative sentiment prediction, Matisse provides fine-grained emotion classification based on Valence, Arousal, and Dominance dimensions and a novel machine learning process. Information from the sentiment/emotion analytics are fused with raw data and summary information to feed temporal, geospatial, term frequency, and scatterplot visualizations using a multi-scale, coordinated interaction model. After describing these techniques, we conclude with a practical case study focused on analyzing the Twitter sample stream during the week of the 2013 Boston Marathon bombings. The case study demonstrates the effectiveness of Matisse at providing guided situational awareness of significant trends in social media streams by orchestrating computational power and human cognition.

  7. Streaming support for data intensive cloud-based sequence analysis.

    Science.gov (United States)

    Issa, Shadi A; Kienzler, Romeo; El-Kalioby, Mohamed; Tonellato, Peter J; Wall, Dennis; Bruggmann, Rémy; Abouelhoda, Mohamed

    2013-01-01

    Cloud computing provides a promising solution to the genomics data deluge problem resulting from the advent of next-generation sequencing (NGS) technology. Based on the concepts of "resources-on-demand" and "pay-as-you-go", scientists with no or limited infrastructure can have access to scalable and cost-effective computational resources. However, the large size of NGS data causes a significant data transfer latency from the client's site to the cloud, which presents a bottleneck for using cloud computing services. In this paper, we provide a streaming-based scheme to overcome this problem, where the NGS data is processed while being transferred to the cloud. Our scheme targets the wide class of NGS data analysis tasks, where the NGS sequences can be processed independently from one another. We also provide the elastream package that supports the use of this scheme with individual analysis programs or with workflow systems. Experiments presented in this paper show that our solution mitigates the effect of data transfer latency and saves both time and cost of computation.

  8. Streaming Support for Data Intensive Cloud-Based Sequence Analysis

    Directory of Open Access Journals (Sweden)

    Shadi A. Issa

    2013-01-01

    Full Text Available Cloud computing provides a promising solution to the genomics data deluge problem resulting from the advent of next-generation sequencing (NGS technology. Based on the concepts of “resources-on-demand” and “pay-as-you-go”, scientists with no or limited infrastructure can have access to scalable and cost-effective computational resources. However, the large size of NGS data causes a significant data transfer latency from the client’s site to the cloud, which presents a bottleneck for using cloud computing services. In this paper, we provide a streaming-based scheme to overcome this problem, where the NGS data is processed while being transferred to the cloud. Our scheme targets the wide class of NGS data analysis tasks, where the NGS sequences can be processed independently from one another. We also provide the elastream package that supports the use of this scheme with individual analysis programs or with workflow systems. Experiments presented in this paper show that our solution mitigates the effect of data transfer latency and saves both time and cost of computation.

  9. Streaming Support for Data Intensive Cloud-Based Sequence Analysis

    Science.gov (United States)

    Issa, Shadi A.; Kienzler, Romeo; El-Kalioby, Mohamed; Tonellato, Peter J.; Wall, Dennis; Bruggmann, Rémy; Abouelhoda, Mohamed

    2013-01-01

    Cloud computing provides a promising solution to the genomics data deluge problem resulting from the advent of next-generation sequencing (NGS) technology. Based on the concepts of “resources-on-demand” and “pay-as-you-go”, scientists with no or limited infrastructure can have access to scalable and cost-effective computational resources. However, the large size of NGS data causes a significant data transfer latency from the client's site to the cloud, which presents a bottleneck for using cloud computing services. In this paper, we provide a streaming-based scheme to overcome this problem, where the NGS data is processed while being transferred to the cloud. Our scheme targets the wide class of NGS data analysis tasks, where the NGS sequences can be processed independently from one another. We also provide the elastream package that supports the use of this scheme with individual analysis programs or with workflow systems. Experiments presented in this paper show that our solution mitigates the effect of data transfer latency and saves both time and cost of computation. PMID:23710461

  10. Hydrogeochemical and stream sediment reconnaissance basic data for Lubbock NTMS Quadrangle, Texas

    International Nuclear Information System (INIS)

    1979-01-01

    Field and laboratory data are presented for 994 groundwater and 602 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided, and pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Interpretation of the groundwater data indicate that the area which appears most promising for uranium mineralization is located in the southwestern part of the quadrangle, particularly in Crosby, Garza, Lynn, and Lubbock Counties. The waters produced from the Ogallala Formation in this area have high values for arsenic, molybdenum, selenium, and vanadium. Groundwaters from the Dockum Group in Garza County where uranium is associated with selenium, molybdenum, and copper indicate potential for uranium mineralization. Uranium is generally associated with copper, iron, and sulfate in the Permian aquifers reflecting the red bed evaporite lithology of those units. The stream sediment data indicate that the Dockum Group has the highest potential for uranium mineralization, particularly in and around Garza County. Associated elements indicate that uranium may occur in residual minerals or in hydrous manganese oxides. Sediment data also indicate that the Blaine Formation shows limited potential for small red bed copper-uranium deposits

  11. Optimal Multi-Interface Selection for Mobile Video Streaming in Efficient Battery Consumption and Data Usage

    Directory of Open Access Journals (Sweden)

    Seonghoon Moon

    2016-01-01

    Full Text Available With the proliferation of high-performance, large-screen mobile devices, users’ expectations of having access to high-resolution video content in smooth network environments are steadily growing. To guarantee such stable streaming, a high cellular network bandwidth is required; yet network providers often charge high prices for even limited data plans. Moreover, the costs of smoothly streaming high-resolution videos are not merely monetary; the device’s battery life must also be accounted for. To resolve these problems, we design an optimal multi-interface selection system for streaming video over HTTP/TCP. An optimization problem including battery life and LTE data constraints is derived and then solved using binary integer programming. Additionally, the system is designed with an adoption of split-layer scalable video coding, which provides direct adaptations of video quality and prevents out-of-order packet delivery problems. The proposed system is evaluated using a prototype application in a real, iOS-based device as well as through experiments conducted in heterogeneous mobile scenarios. Results show that the system not only guarantees the highest-possible video quality, but also prevents reckless consumption of LTE data and battery life.

  12. A survey of systems for massive stream analytics

    OpenAIRE

    Singh, Maninder Pal; Hoque, Mohammad A.; Tarkoma, Sasu

    2016-01-01

    The immense growth of data demands switching from traditional data processing solutions to systems, which can process a continuous stream of real time data. Various applications employ stream processing systems to provide solutions to emerging Big Data problems. Open-source solutions such as Storm, Spark Streaming, and S4 are the attempts to answer key stream processing questions. The recent introduction of real time stream processing commercial solutions such as Amazon Kinesis, IBM Infospher...

  13. Rapid Bioassessment Methods for Assessing Stream Macroinvertebrate Community on the Savannah River Site

    International Nuclear Information System (INIS)

    Specht, W.L.

    1999-01-01

    Macroinvertebrate sampling was performed at 16 locations in the Savannah River Site (SRS) streams using Hester-Dendy multiplate samplers and EPA Rapid Bioassessment Protocols (RBP). Some of the sampling locations were unimpacted, while other locations had been subject to various forms of perturbation by SRS activities. In general, the data from the Hester-Dendy multiplate samplers were more sensitive at detecting impacts than were the RBP data. We developed a Biotic Index for the Hester-Dendy data which incorporated eight community structure, function, and balance parameters. when tested using a data set that was unrelated to the data set that was used in developing the Biotic Index, the index was very successful at detecting impact

  14. Rapid Bioassessment Methods for Assessing Stream Macroinvertebrate Community on the Savannah River Site

    Energy Technology Data Exchange (ETDEWEB)

    Specht, W.L.

    1999-11-22

    Macroinvertebrate sampling was performed at 16 locations in the Savannah River Site (SRS) streams using Hester-Dendy multiplate samplers and EPA Rapid Bioassessment Protocols (RBP). Some of the sampling locations were unimpacted, while other locations had been subject to various forms of perturbation by SRS activities. In general, the data from the Hester-Dendy multiplate samplers were more sensitive at detecting impacts than were the RBP data. We developed a Biotic Index for the Hester-Dendy data which incorporated eight community structure, function, and balance parameters. when tested using a data set that was unrelated to the data set that was used in developing the Biotic Index, the index was very successful at detecting impact.

  15. Dynamical modeling of tidal streams

    International Nuclear Information System (INIS)

    Bovy, Jo

    2014-01-01

    I present a new framework for modeling the dynamics of tidal streams. The framework consists of simple models for the initial action-angle distribution of tidal debris, which can be straightforwardly evolved forward in time. Taking advantage of the essentially one-dimensional nature of tidal streams, the transformation to position-velocity coordinates can be linearized and interpolated near a small number of points along the stream, thus allowing for efficient computations of a stream's properties in observable quantities. I illustrate how to calculate the stream's average location (its 'track') in different coordinate systems, how to quickly estimate the dispersion around its track, and how to draw mock stream data. As a generative model, this framework allows one to compute the full probability distribution function and marginalize over or condition it on certain phase-space dimensions as well as convolve it with observational uncertainties. This will be instrumental in proper data analysis of stream data. In addition to providing a computationally efficient practical tool for modeling the dynamics of tidal streams, the action-angle nature of the framework helps elucidate how the observed width of the stream relates to the velocity dispersion or mass of the progenitor, and how the progenitors of 'orphan' streams could be located. The practical usefulness of the proposed framework crucially depends on the ability to calculate action-angle variables for any orbit in any gravitational potential. A novel method for calculating actions, frequencies, and angles in any static potential using a single orbit integration is described in the Appendix.

  16. Time course of auditory streaming: Do CI users differ from normal-hearing listeners?

    Directory of Open Access Journals (Sweden)

    Martin eBöckmann-Barthel

    2014-07-01

    Full Text Available In a complex acoustical environment with multiple sound sources the auditory system uses streaming as a tool to organize the incoming sounds in one or more streams depending on the stimulus parameters. Streaming is commonly studied by alternating sequences of signals. These are often tones with different frequencies. The present study investigates stream segregation in cochlear implant (CI users, where hearing is restored by electrical stimulation of the auditory nerve. CI users listened to 30-s long sequences of alternating A and B harmonic complexes at four different fundamental frequency separations, ranging from 2 to 14 semitones. They had to indicate as promptly as possible after sequence onset, if they perceived one stream or two streams and, in addition, any changes of the percept throughout the rest of the sequence. The conventional view is that the initial percept is always that of a single stream which may after some time change to a percept of two streams. This general build-up hypothesis has recently been challenged on the basis of a new analysis of data of normal-hearing listeners which showed a build-up response only for an intermediate frequency separation. Using the same experimental paradigm and analysis, the present study found that the results of CI users agree with those of the normal-hearing listeners: (i the probability of the first decision to be a one-stream percept decreased and that of a two-stream percept increased as Δf increased, and (ii a build-up was only found for 6 semitones. Only the time elapsed before the listeners made their first decision of the percept was prolonged as compared to normal-hearing listeners. The similarity in the data of the CI user and the normal-hearing listeners indicates that the quality of stream formation is similar in these groups of listeners.

  17. Multi-scale window specification over streaming trajectories

    Directory of Open Access Journals (Sweden)

    Kostas Patroumpas

    2013-12-01

    Full Text Available Enormous amounts of positional information are collected by monitoring applications in domains such as fleet management, cargo transport, wildlife protection, etc. With the advent of modern location-based services, processing such data mostly focuses on providing real-time response to a variety of user requests in continuous and scalable fashion. An important class of such queries concerns evolving trajectories that continuously trace the streaming locations of moving objects, like GPS-equipped vehicles, commodities with RFID's, people with smartphones etc. In this work, we propose an advanced windowing operator that enables online, incremental examination of recent motion paths at multiple resolutions for numerous point entities. When applied against incoming positions, this window can abstract trajectories at coarser representations towards the past, while retaining progressively finer features closer to the present. We explain the semantics of such multi-scale sliding windows through parameterized functions that reflect the sequential nature of trajectories and can effectively capture their spatiotemporal properties. Such window specification goes beyond its usual role for non-blocking processing of multiple concurrent queries. Actually, it can offer concrete subsequences from each trajectory, thus preserving continuity in time and contiguity in space along the respective segments. Further, we suggest language extensions in order to express characteristic spatiotemporal queries using windows. Finally, we discuss algorithms for nested maintenance of multi-scale windows and evaluate their efficiency against streaming positional data, offering empirical evidence of their benefits to online trajectory processing.

  18. Estimating Discharge and Nonpoint Source Nitrate Loading to Streams From Three End-Member Pathways Using High-Frequency Water Quality Data

    Science.gov (United States)

    Miller, Matthew P.; Tesoriero, Anthony J.; Hood, Krista; Terziotti, Silvia; Wolock, David M.

    2017-12-01

    The myriad hydrologic and biogeochemical processes taking place in watersheds occurring across space and time are integrated and reflected in the quantity and quality of water in streams and rivers. Collection of high-frequency water quality data with sensors in surface waters provides new opportunities to disentangle these processes and quantify sources and transport of water and solutes in the coupled groundwater-surface water system. A new approach for separating the streamflow hydrograph into three components was developed and coupled with high-frequency nitrate data to estimate time-variable nitrate loads from chemically dilute quick flow, chemically concentrated quick flow, and slowflow groundwater end-member pathways for periods of up to 2 years in a groundwater-dominated and a quick-flow-dominated stream in central Wisconsin, using only streamflow and in-stream water quality data. The dilute and concentrated quick flow end-members were distinguished using high-frequency specific conductance data. Results indicate that dilute quick flow contributed less than 5% of the nitrate load at both sites, whereas 89 ± 8% of the nitrate load at the groundwater-dominated stream was from slowflow groundwater, and 84 ± 25% of the nitrate load at the quick-flow-dominated stream was from concentrated quick flow. Concentrated quick flow nitrate concentrations varied seasonally at both sites, with peak concentrations in the winter that were 2-3 times greater than minimum concentrations during the growing season. Application of this approach provides an opportunity to assess stream vulnerability to nonpoint source nitrate loading and expected stream responses to current or changing conditions and practices in watersheds.

  19. Real-Time Earthquake Intensity Estimation Using Streaming Data Analysis of Social and Physical Sensors

    Science.gov (United States)

    Kropivnitskaya, Yelena; Tiampo, Kristy F.; Qin, Jinhui; Bauer, Michael A.

    2017-06-01

    Earthquake intensity is one of the key components of the decision-making process for disaster response and emergency services. Accurate and rapid intensity calculations can help to reduce total loss and the number of casualties after an earthquake. Modern intensity assessment procedures handle a variety of information sources, which can be divided into two main categories. The first type of data is that derived from physical sensors, such as seismographs and accelerometers, while the second type consists of data obtained from social sensors, such as witness observations of the consequences of the earthquake itself. Estimation approaches using additional data sources or that combine sources from both data types tend to increase intensity uncertainty due to human factors and inadequate procedures for temporal and spatial estimation, resulting in precision errors in both time and space. Here we present a processing approach for the real-time analysis of streams of data from both source types. The physical sensor data is acquired from the U.S. Geological Survey (USGS) seismic network in California and the social sensor data is based on Twitter user observations. First, empirical relationships between tweet rate and observed Modified Mercalli Intensity (MMI) are developed using data from the M6.0 South Napa, CAF earthquake that occurred on August 24, 2014. Second, the streams of both data types are analyzed together in simulated real-time to produce one intensity map. The second implementation is based on IBM InfoSphere Streams, a cloud platform for real-time analytics of big data. To handle large processing workloads for data from various sources, it is deployed and run on a cloud-based cluster of virtual machines. We compare the quality and evolution of intensity maps from different data sources over 10-min time intervals immediately following the earthquake. Results from the joint analysis shows that it provides more complete coverage, with better accuracy and higher

  20. Passing a smoke-free law in a pro-tobacco culture: a multiple streams approach.

    Science.gov (United States)

    Greathouse, Lisa W; Hahn, Ellen J; Okoli, Chizimuzo T C; Warnick, Todd A; Riker, Carol A

    2005-08-01

    This article describes a case study of the policy development and political decision-making process involved in the enactment of Lexington, Kentucky's smoke-free law. The multiple streams framework is used to analyze the development of the law in a seemingly unlikely and challenging political environment. Proponents developed a dissemination research plan targeted at policy makers and the public to demonstrate the need for a comprehensive law. The existence of a strong coalition of health care providers and health care systems including the board of health, as well as long-standing tobacco control expertise and a strong legal team, were essential ingredients for success. A deliberate strategy to expose the tobacco industry was effective in preparing policy makers for the opponents' policy arguments. As expected, a hospitality industry association was formed to oppose the ordinance, resulting in a legal challenge that delayed enactment of the law.

  1. Automatic event-based synchronization of multimodal data streams from wearable and ambient sensors

    NARCIS (Netherlands)

    Bannach, D.; Amft, O.D.; Lukowicz, P.; Barnaghi, P.; Moessner, K.; Presser, M.; Meissner, S.

    2009-01-01

    A major challenge in using multi-modal, distributed sensor systems for activity recognition is to maintain a temporal synchronization between individually recorded data streams. A common approach is to use well defined ‘synchronization actions’ performed by the user to generate, easily identifiable

  2. Stream Tracker: Crowd sourcing and remote sensing to monitor stream flow intermittence

    Science.gov (United States)

    Puntenney, K.; Kampf, S. K.; Newman, G.; Lefsky, M. A.; Weber, R.; Gerlich, J.

    2017-12-01

    Streams that do not flow continuously in time and space support diverse aquatic life and can be critical contributors to downstream water supply. However, these intermittent streams are rarely monitored and poorly mapped. Stream Tracker is a community powered stream monitoring project that pairs citizen contributed observations of streamflow presence or absence with a network of streamflow sensors and remotely sensed data from satellites to track when and where water is flowing in intermittent stream channels. Citizens can visit sites on roads and trails to track flow and contribute their observations to the project site hosted by CitSci.org. Data can be entered using either a mobile application with offline capabilities or an online data entry portal. The sensor network provides a consistent record of streamflow and flow presence/absence across a range of elevations and drainage areas. Capacitance, resistance, and laser sensors have been deployed to determine the most reliable, low cost sensor that could be mass distributed to track streamflow intermittence over a larger number of sites. Streamflow presence or absence observations from the citizen and sensor networks are then compared to satellite imagery to improve flow detection algorithms using remotely sensed data from Landsat. In the first two months of this project, 1,287 observations have been made at 241 sites by 24 project members across northern and western Colorado.

  3. Phase II in-stream sediment measurements in the Talakhaya watershed on Rota, Commonwealth of the Northern Mariana Islands from 2016-03-09 to 2017-05-11 (NCEI Accession 0166379)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The dataset includes TSS (mg/L), turbidity (NTU), stream level (ft), and precipitation (in) data collected at multiple stations across five streams and at four rain...

  4. Use of NTRIP for Optimizing the Decoding Algorithm for Real-Time Data Streams

    Directory of Open Access Journals (Sweden)

    Zhanke He

    2014-10-01

    Full Text Available As a network transmission protocol, Networked Transport of RTCM via Internet Protocol (NTRIP is widely used in GPS and Global Orbiting Navigational Satellite System (GLONASS Augmentation systems, such as Continuous Operational Reference System (CORS, Wide Area Augmentation System (WAAS and Satellite Based Augmentation Systems (SBAS. With the deployment of BeiDou Navigation Satellite system(BDS to serve the Asia-Pacific region, there are increasing needs for ground monitoring of the BeiDou Navigation Satellite system and the development of the high-precision real-time BeiDou products. This paper aims to optimize the decoding algorithm of NTRIP Client data streams and the user authentication strategies of the NTRIP Caster based on NTRIP. The proposed method greatly enhances the handling efficiency and significantly reduces the data transmission delay compared with the Federal Agency for Cartography and Geodesy (BKG NTRIP. Meanwhile, a transcoding method is proposed to facilitate the data transformation from the BINary EXchange (BINEX format to the RTCM format. The transformation scheme thus solves the problem of handing real-time data streams from Trimble receivers in the BeiDou Navigation Satellite System indigenously developed by China.

  5. Use of NTRIP for optimizing the decoding algorithm for real-time data streams.

    Science.gov (United States)

    He, Zhanke; Tang, Wenda; Yang, Xuhai; Wang, Liming; Liu, Jihua

    2014-10-10

    As a network transmission protocol, Networked Transport of RTCM via Internet Protocol (NTRIP) is widely used in GPS and Global Orbiting Navigational Satellite System (GLONASS) Augmentation systems, such as Continuous Operational Reference System (CORS), Wide Area Augmentation System (WAAS) and Satellite Based Augmentation Systems (SBAS). With the deployment of BeiDou Navigation Satellite system(BDS) to serve the Asia-Pacific region, there are increasing needs for ground monitoring of the BeiDou Navigation Satellite system and the development of the high-precision real-time BeiDou products. This paper aims to optimize the decoding algorithm of NTRIP Client data streams and the user authentication strategies of the NTRIP Caster based on NTRIP. The proposed method greatly enhances the handling efficiency and significantly reduces the data transmission delay compared with the Federal Agency for Cartography and Geodesy (BKG) NTRIP. Meanwhile, a transcoding method is proposed to facilitate the data transformation from the BINary EXchange (BINEX) format to the RTCM format. The transformation scheme thus solves the problem of handing real-time data streams from Trimble receivers in the BeiDou Navigation Satellite System indigenously developed by China.

  6. The Midwest Stream Quality Assessment—Influences of human activities on streams

    Science.gov (United States)

    Van Metre, Peter C.; Mahler, Barbara J.; Carlisle, Daren M.; Coles, James F.

    2018-04-16

    Healthy streams and the fish and other organisms that live in them contribute to our quality of life. Extensive modification of the landscape in the Midwestern United States, however, has profoundly affected the condition of streams. Row crops and pavement have replaced grasslands and woodlands, streams have been straightened, and wetlands and fields have been drained. Runoff from agricultural and urban land brings sediment and chemicals to streams. What is the chemical, physical, and biological condition of Midwestern streams? Which physical and chemical stressors are adversely affecting biological communities, what are their origins, and how might we lessen or avoid their adverse effects?In 2013, the U.S. Geological Survey (USGS) conducted the Midwest Stream Quality Assessment to evaluate how human activities affect the biological condition of Midwestern streams. In collaboration with the U.S. Environmental Protection Agency National Rivers and Streams Assessment, the USGS sampled 100 streams, chosen to be representative of the different types of watersheds in the region. Biological condition was evaluated based on the number and diversity of fish, algae, and invertebrates in the streams. Changes to the physical habitat and chemical characteristics of the streams—“stressors”—were assessed, and their relation to landscape factors and biological condition was explored by using mathematical models. The data and models help us to better understand how the human activities on the landscape are affecting streams in the region.

  7. The Stream-Catchment (StreamCat) Dataset: A database of watershed metrics for the conterminous USA

    Science.gov (United States)

    We developed an extensive database of landscape metrics for ~2.65 million streams, and their associated catchments, within the conterminous USA: The Stream-Catchment (StreamCat) Dataset. These data are publically available and greatly reduce the specialized geospatial expertise n...

  8. Results of Macroinvertebrate Sampling Conducted at 33 SRS Stream Locations, July--August 1993

    Energy Technology Data Exchange (ETDEWEB)

    Specht, W.L.

    1994-12-01

    In order to assess the health of the macroinvertebrate communities of SRS streams, the macroinvertebrate communities at 30 stream locations on SRS were sampled during the summer of 1993, using Hester-Dendy multiplate samplers. In addition, three off-site locations in the Upper Three Runs drainage were sampled in order to assess the potential for impact from off-site activities. In interpreting the data, it is important to recognize that these data were from a single set of collections. Macroinvertebrate communities often undergo considerable temporal variation, and are also greatly influenced by such factors as water depth, water velocity, and available habitat. These stations were selected with the intent of developing an on-going sampling program at a smaller number of stations, with the selection of the stations to be based largely upon the results of this preliminary sampling program. When stations within a given stream showed similar results, fewer stations would be sampled in the future. Similarly, if a stream appeared to be perturbed, additional stations or chemical analyses might be added so that the source of the perturbation could be identified. In general, unperturbed streams will contain more taxa than perturbed streams, and the distribution of taxa among orders or families will differ. Some groups of macroinvertebrates, such as Ephemeroptera (mayflies), Plecoptera (stoneflies) and Trichoptera (caddisflies), which are collectively called EPT taxa, are considered to be relatively sensitive to most kinds of stream perturbation; therefore a reduced number of EPT taxa generally indicates that the stream has been subject to chemical or physical stressors. In coastal plain streams, EPT taxa are generally less dominant than in streams with rocky substrates, while Chironomidae (midges) are more abundant. (Abstract Truncated)

  9. A decision support system using combined-classifier for high-speed data stream in smart grid

    Science.gov (United States)

    Yang, Hang; Li, Peng; He, Zhian; Guo, Xiaobin; Fong, Simon; Chen, Huajun

    2016-11-01

    Large volume of high-speed streaming data is generated by big power grids continuously. In order to detect and avoid power grid failure, decision support systems (DSSs) are commonly adopted in power grid enterprises. Among all the decision-making algorithms, incremental decision tree is the most widely used one. In this paper, we propose a combined classifier that is a composite of a cache-based classifier (CBC) and a main tree classifier (MTC). We integrate this classifier into a stream processing engine on top of the DSS such that high-speed steaming data can be transformed into operational intelligence efficiently. Experimental results show that our proposed classifier can return more accurate answers than other existing ones.

  10. Linking Stream Dissolved Oxygen with the Dynamic Environmental Drivers across the Pacific Coast of U.S.A.

    Science.gov (United States)

    Araya, F. Z.; Abdul-Aziz, O. I.

    2017-12-01

    This study utilized a systematic data analytics approach to determine the relative linkages of stream dissolved oxygen (DO) with the hydro-climatic and biogeochemical drivers across the U.S. Pacific Coast. Multivariate statistical techniques of Pearson correlation matrix, principal component analysis, and factor analysis were applied to a complex water quality dataset (1998-2015) at 35 water quality monitoring stations of USGS NWIS and EPA STORET. Power-law based partial least squares regression (PLSR) models with a bootstrap Monte Carlo procedure (1000 iterations) were developed to reliably estimate the relative linkages by resolving multicollinearity (Nash-Sutcliffe Efficiency, NSE = 0.50-0.94). Based on the dominant drivers, four environmental regimes have been identified and adequately described the system-data variances. In Pacific North West and Southern California, water temperature was the most dominant driver of DO in majority of the streams. However, in Central and Northern California, stream DO was controlled by multiple drivers (i.e., water temperature, pH, stream flow, and total phosphorus), exhibiting a transitional environmental regime. Further, total phosphorus (TP) appeared to be the limiting nutrient for most streams. The estimated linkages and insights would be useful to identify management priorities to achieve healthy coastal stream ecosystems across the Pacific Coast of U.S.A. and similar regions around the world. Keywords: Data analytics, water quality, coastal streams, dissolved oxygen, environmental regimes, Pacific Coast, United States.

  11. Development of Ecogeomorphological (EGM Stream Design and Assessment Tools for the Piedmont of Alabama, USA

    Directory of Open Access Journals (Sweden)

    Brian Helms

    2016-04-01

    Full Text Available Regional data needed for effective stream restoration include hydraulic geometry relationships (i.e., regional curves and reference channel morphology parameters. Increasingly ecological conditions are being considered when designing, implementing, and assessing restoration efforts. We provide morphology relationships and associated ecological endpoint curves for reference streams in the Alabama piedmont. Twenty-one reference stream reaches were identified in the Tallapoosa drainage of Alabama, ranging from 0.2 to 242 km2 drainage area. Geomorphic surveys were conducted in each stream to measure riffle cross-sections and longitudinal profiles and related to drainage area to develop regional curves. Fish, crayfish, and benthic macroinvertebrates were collected from each surveyed reach and related to drainage area and geomorphic data to provide associated biological community endpoints. Bankfull channel cross-section area, width, mean depth, and estimated discharge were strongly correlated to watershed drainage area, similar to efforts in other areas of the Piedmont ecoregion. Multiple measures of fish assemblages and crayfish size were strongly predicted by drainage area and geomorphic dimensions. Macroinvertebrates showed no taxonomic and limited functional relationships with drainage area and geomorphic dimension. These tools, which integrate geomorphological and ecological conditions, can result in improved stream evaluations and designs increasing the effectiveness of stream restoration projects.

  12. Precipitation and stream water stable isotope data from the Marys River, Oregon in water year 2015.

    Data.gov (United States)

    U.S. Environmental Protection Agency — Water stable isotope data collected from a range of streams throughout the Marys River basin in water year 2015, and precipitation data collected within the basin at...

  13. Extending Data Worth Analyses to Select Multiple Observations Targeting Multiple Forecasts

    DEFF Research Database (Denmark)

    Vilhelmsen, Troels Norvin; Ferre, Ty Paul

    2017-01-01

    . In the present study, we extend previous data worth analyses to include: simultaneous selection of multiple new measurements and consideration of multiple forecasts of interest. We show how the suggested approach can be used to optimize data collection. This can be used in a manner that suggests specific...... measurement sets or that produces probability maps indicating areas likely to be informative for specific forecasts. Moreover, we provide examples documenting that sequential measurement election approaches often lead to suboptimal designs and that estimates of data covariance should be included when...

  14. Multiple drivers, scales, and interactions influence southern Appalachian stream salamander occupancy

    Science.gov (United States)

    Cecala, Kristen K.; Maerz, John C.; Halstead, Brian J.; Frisch, John R.; Gragson, Ted L.; Hepinstall-Cymerman, Jeffrey; Leigh, David S.; Jackson, C. Rhett; Peterson, James T.; Pringle, Catherine M.

    2018-01-01

    Understanding how factors that vary in spatial scale relate to population abundance is vital to forecasting species responses to environmental change. Stream and river ecosystems are inherently hierarchical, potentially resulting in organismal responses to fine‐scale changes in patch characteristics that are conditional on the watershed context. Here, we address how populations of two salamander species are affected by interactions among hierarchical processes operating at different scales within a rapidly changing landscape of the southern Appalachian Mountains. We modeled reach‐level occupancy of larval and adult black‐bellied salamanders (Desmognathus quadramaculatus) and larval Blue Ridge two‐lined salamanders (Eurycea wilderae) as a function of 17 different terrestrial and aquatic predictor variables that varied in spatial extent. We found that salamander occurrence varied widely among streams within fully forested catchments, but also exhibited species‐specific responses to changes in local conditions. While D. quadramaculatus declined predictably in relation to losses in forest cover, larval occupancy exhibited the strongest negative response to forest loss as well as decreases in elevation. Conversely, occupancy of E. wilderae was unassociated with watershed conditions, only responding negatively to higher proportions of fast‐flowing stream habitat types. Evaluation of hierarchical relationships demonstrated that most fine‐scale variables were closely correlated with broad watershed‐scale variables, suggesting that local reach‐scale factors have relatively smaller effects within the context of the larger landscape. Our results imply that effective management of southern Appalachian stream salamanders must first focus on the larger scale condition of watersheds before management of local‐scale conditions should proceed. Our findings confirm the results of some studies while refuting the results of others, which may indicate that

  15. Towards Cache-Enabled, Order-Aware, Ontology-Based Stream Reasoning Framework

    Energy Technology Data Exchange (ETDEWEB)

    Yan, Rui; Praggastis, Brenda L.; Smith, William P.; McGuinness, Deborah L.

    2016-08-16

    While streaming data have become increasingly more popular in business and research communities, semantic models and processing software for streaming data have not kept pace. Traditional semantic solutions have not addressed transient data streams. Semantic web languages (e.g., RDF, OWL) have typically addressed static data settings and linked data approaches have predominantly addressed static or growing data repositories. Streaming data settings have some fundamental differences; in particular, data are consumed on the fly and data may expire. Stream reasoning, a combination of stream processing and semantic reasoning, has emerged with the vision of providing "smart" processing of streaming data. C-SPARQL is a prominent stream reasoning system that handles semantic (RDF) data streams. Many stream reasoning systems including C-SPARQL use a sliding window and use data arrival time to evict data. For data streams that include expiration times, a simple arrival time scheme is inadequate if the window size does not match the expiration period. In this paper, we propose a cache-enabled, order-aware, ontology-based stream reasoning framework. This framework consumes RDF streams with expiration timestamps assigned by the streaming source. Our framework utilizes both arrival and expiration timestamps in its cache eviction policies. In addition, we introduce the notion of "semantic importance" which aims to address the relevance of data to the expected reasoning, thus enabling the eviction algorithms to be more context- and reasoning-aware when choosing what data to maintain for question answering. We evaluate this framework by implementing three different prototypes and utilizing five metrics. The trade-offs of deploying the proposed framework are also discussed.

  16. RStorm: Developing and Testing Streaming Algorithms in R

    NARCIS (Netherlands)

    Kaptein, M.C.

    2014-01-01

    Streaming data, consisting of indefinitely evolving sequences, are becoming ubiquitous in many branches of science and in various applications. Computer scientists have developed streaming applications such as Storm and the S4 distributed stream computing platform1 to deal with data streams.

  17. RStorm : Developing and testing streaming algorithms in R

    NARCIS (Netherlands)

    Kaptein, M.C.

    2014-01-01

    Streaming data, consisting of indefinitely evolving sequences, are becoming ubiquitous in many branches of science and in various applications. Computer scientists have developed streaming applications such as Storm and the S4 distributed stream computing platform1 to deal with data streams.

  18. Multiple-Features-Based Semisupervised Clustering DDoS Detection Method

    Directory of Open Access Journals (Sweden)

    Yonghao Gu

    2017-01-01

    Full Text Available DDoS attack stream from different agent host converged at victim host will become very large, which will lead to system halt or network congestion. Therefore, it is necessary to propose an effective method to detect the DDoS attack behavior from the massive data stream. In order to solve the problem that large numbers of labeled data are not provided in supervised learning method, and the relatively low detection accuracy and convergence speed of unsupervised k-means algorithm, this paper presents a semisupervised clustering detection method using multiple features. In this detection method, we firstly select three features according to the characteristics of DDoS attacks to form detection feature vector. Then, Multiple-Features-Based Constrained-K-Means (MF-CKM algorithm is proposed based on semisupervised clustering. Finally, using MIT Laboratory Scenario (DDoS 1.0 data set, we verify that the proposed method can improve the convergence speed and accuracy of the algorithm under the condition of using a small amount of labeled data sets.

  19. A novel WDM passive optical network architecture supporting two independent multicast data streams

    Science.gov (United States)

    Qiu, Yang; Chan, Chun-Kit

    2012-01-01

    We propose a novel scheme to perform optical multicast overlay of two independent multicast data streams on a wavelength-division-multiplexed (WDM) passive optical network. By controlling a sinusoidal clock signal and shifting the wavelength at the optical line terminal (OLT), the delivery of the two multicast data, being carried by the generated optical tones, can be independently and flexibly controlled. Simultaneous transmission of 10-Gb/s unicast downstream and upstream data as well as two independent 10-Gb/s multicast data was successfully demonstrated.

  20. MIVIS image geocoding experience on merging position attitude system data and public domain GPS stream (ASI-GeoDAF

    Directory of Open Access Journals (Sweden)

    S. Pignatti

    2006-06-01

    Full Text Available The use of airborne scanners involves geo-referencing problems, which are difficult because of the need to know the exact platform position and attitude for each scan line. The errors of the onboard navigation system are normally corrected using ground control point on the image. This post-processing correction procedure is too long in case of multiple flight campaigns, and besides it implies the need to have available 1:10000 orthophotoimages or maps in digital format. To optimize the above procedure a new method to correct MIVIS navigational data in the post-processing phase has been implemented. The procedure takes into consideration the GPS stream in Rinex format of common knowledge and findable on the web, acquired at the ground stations of the Geodetic Data Archiving Facilities provided by ASI. The application of this correction entails the assumption that the environmental variables affecting both onboard and geodetic GPS equally affect the position measurements. The airborne data correction was carried out merging the two data sets (onboard and ground station GPS to achieve a more precise aircraft trajectory. The present study compares the geo-coded images obtained by means of the two post-processing methods.

  1. Estimating stream discharge from a Himalayan Glacier using coupled satellite sensor data

    Science.gov (United States)

    Child, S. F.; Stearns, L. A.; van der Veen, C. J.; Haritashya, U. K.; Tarpanelli, A.

    2015-12-01

    The 4th IPCC report highlighted our limited understanding of Himalayan glacier behavior and contribution to the region's hydrology. Seasonal snow and glacier melt in the Himalayas are important sources of water, but estimates greatly differ about the actual contribution of melted glacier ice to stream discharge. A more comprehensive understanding of the contribution of glaciers to stream discharge is needed because streams being fed by glaciers affect the livelihoods of a large part of the world's population. Most of the streams in the Himalayas are unmonitored because in situ measurements are logistically difficult and costly. This necessitates the use of remote sensing platforms to obtain estimates of river discharge for validating hydrological models. In this study, we estimate stream discharge using cost-effective methods via repeat satellite imagery from Landsat-8 and SENTINEL-1A sensors. The methodology is based on previous studies, which show that ratio values from optical satellite bands correlate well with measured stream discharge. While similar, our methodology relies on significantly higher resolution imagery (30 m) and utilizes bands that are in the blue and near-infrared spectrum as opposed to previous studies using 250 m resolution imagery and spectral bands only in the near-infrared. Higher resolution imagery is necessary for streams where the source is a glacier's terminus because the width of the stream is often only 10s of meters. We validate our methodology using two rivers in the state of Kansas, where stream gauges are plentiful. We then apply our method to the Bhagirathi River, in the North-Central Himalayas, which is fed by the Gangotri Glacier and has a well monitored stream gauge. The analysis will later be used to couple river discharge and glacier flow and mass balance through an integrated hydrologic model in the Bhagirathi Basin.

  2. Tryton Supercomputer Capabilities for Analysis of Massive Data Streams

    Directory of Open Access Journals (Sweden)

    Krawczyk Henryk

    2015-09-01

    Full Text Available The recently deployed supercomputer Tryton, located in the Academic Computer Center of Gdansk University of Technology, provides great means for massive parallel processing. Moreover, the status of the Center as one of the main network nodes in the PIONIER network enables the fast and reliable transfer of data produced by miscellaneous devices scattered in the area of the whole country. The typical examples of such data are streams containing radio-telescope and satellite observations. Their analysis, especially with real-time constraints, can be challenging and requires the usage of dedicated software components. We propose a solution for such parallel analysis using the supercomputer, supervised by the KASKADA platform, which with the conjunction with immerse 3D visualization techniques can be used to solve problems such as pulsar detection and chronometric or oil-spill simulation on the sea surface.

  3. Developing semi-analytical solution for multiple-zone transient storage model with spatially non-uniform storage

    Science.gov (United States)

    Deng, Baoqing; Si, Yinbing; Wang, Jia

    2017-12-01

    Transient storages may vary along the stream due to stream hydraulic conditions and the characteristics of storage. Analytical solutions of transient storage models in literature didn't cover the spatially non-uniform storage. A novel integral transform strategy is presented that simultaneously performs integral transforms to the concentrations in the stream and in storage zones by using the single set of eigenfunctions derived from the advection-diffusion equation of the stream. The semi-analytical solution of the multiple-zone transient storage model with the spatially non-uniform storage is obtained by applying the generalized integral transform technique to all partial differential equations in the multiple-zone transient storage model. The derived semi-analytical solution is validated against the field data in literature. Good agreement between the computed data and the field data is obtained. Some illustrative examples are formulated to demonstrate the applications of the present solution. It is shown that solute transport can be greatly affected by the variation of mass exchange coefficient and the ratio of cross-sectional areas. When the ratio of cross-sectional areas is big or the mass exchange coefficient is small, more reaches are recommended to calibrate the parameter.

  4. Multiple attenuation to reflection seismic data using Radon filter and Wave Equation Multiple Rejection (WEMR) method

    Energy Technology Data Exchange (ETDEWEB)

    Erlangga, Mokhammad Puput [Geophysical Engineering, Institut Teknologi Bandung, Ganesha Street no.10 Basic Science B Buliding fl.2-3 Bandung, 40132, West Java Indonesia puput.erlangga@gmail.com (Indonesia)

    2015-04-16

    Separation between signal and noise, incoherent or coherent, is important in seismic data processing. Although we have processed the seismic data, the coherent noise is still mixing with the primary signal. Multiple reflections are a kind of coherent noise. In this research, we processed seismic data to attenuate multiple reflections in the both synthetic and real seismic data of Mentawai. There are several methods to attenuate multiple reflection, one of them is Radon filter method that discriminates between primary reflection and multiple reflection in the τ-p domain based on move out difference between primary reflection and multiple reflection. However, in case where the move out difference is too small, the Radon filter method is not enough to attenuate the multiple reflections. The Radon filter also produces the artifacts on the gathers data. Except the Radon filter method, we also use the Wave Equation Multiple Elimination (WEMR) method to attenuate the long period multiple reflection. The WEMR method can attenuate the long period multiple reflection based on wave equation inversion. Refer to the inversion of wave equation and the magnitude of the seismic wave amplitude that observed on the free surface, we get the water bottom reflectivity which is used to eliminate the multiple reflections. The WEMR method does not depend on the move out difference to attenuate the long period multiple reflection. Therefore, the WEMR method can be applied to the seismic data which has small move out difference as the Mentawai seismic data. The small move out difference on the Mentawai seismic data is caused by the restrictiveness of far offset, which is only 705 meter. We compared the real free multiple stacking data after processing with Radon filter and WEMR process. The conclusion is the WEMR method can more attenuate the long period multiple reflection than the Radon filter method on the real (Mentawai) seismic data.

  5. Performance Evaluation of Concurrent Multipath Video Streaming in Multihomed Mobile Networks

    Directory of Open Access Journals (Sweden)

    James Nightingale

    2013-01-01

    Full Text Available High-quality real-time video streaming to users in mobile networks is challenging due to the dynamically changing nature of the network paths, particularly the limited bandwidth and varying end-to-end delay. In this paper, we empirically investigate the performance of multipath streaming in the context of multihomed mobile networks. Existing schemes that make use of the aggregated bandwidth of multiple paths can overcome bandwidth limitations on a single path but suffer an efficiency penalty caused by retransmission of lost packets in reliable transport schemes or path switching overheads in unreliable transport schemes. This work focuses on the evaluation of schemes to permit concurrent use of multiple paths to deliver video streams. A comprehensive streaming framework for concurrent multipath video streaming is proposed and experimentally evaluated, using current state-of-the-art H.264 Scalable Video Coding (H.264/SVC and the next generation High Efficiency Video Coding (HEVC standards. It provides a valuable insight into the benefit of using such schemes in conjunction with encoder specific packet prioritisation mechanisms for quality-aware packet scheduling and scalable streaming. The remaining obstacles to deployment of concurrent multipath schemes are identified, and the challenges in realising HEVC based concurrent multipath streaming are highlighted.

  6. StreamAR: incremental and active learning with evolving sensory data for activity recognition

    OpenAIRE

    Abdallah, Z.; Gaber, M.; Srinivasan, B.; Krishnaswamy, S.

    2012-01-01

    Activity recognition focuses on inferring current user activities by leveraging sensory data available on today’s sensor rich environment. Supervised learning has been applied pervasively for activity recognition. Typical activity recognition techniques process sensory data based on point-by-point approaches. In this paper, we propose a novel cluster-based classification for activity recognition Systems, termed StreamAR. The system incorporates incremental and active learning for mining user ...

  7. The role of the geophysical template and environmental regimes in controlling stream-living trout populations

    Science.gov (United States)

    Penaluna, Brooke E.; Railsback, Steve F.; Dunham, Jason B.; Johnson, S.; Bilby, Richard E.; Skaugset, Arne E.

    2015-01-01

    The importance of multiple processes and instream factors to aquatic biota has been explored extensively, but questions remain about how local spatiotemporal variability of aquatic biota is tied to environmental regimes and the geophysical template of streams. We used an individual-based trout model to explore the relative role of the geophysical template versus environmental regimes on biomass of trout (Oncorhynchus clarkii clarkii). We parameterized the model with observed data from each of the four headwater streams (their local geophysical template and environmental regime) and then ran 12 simulations where we replaced environmental regimes (stream temperature, flow, turbidity) of a given stream with values from each neighboring stream while keeping the geophysical template fixed. We also performed single-parameter sensitivity analyses on the model results from each of the four streams. Although our modeled findings show that trout biomass is most responsive to changes in the geophysical template of streams, they also reveal that biomass is restricted by available habitat during seasonal low flow, which is a product of both the stream’s geophysical template and flow regime. Our modeled results suggest that differences in the geophysical template among streams render trout more or less sensitive to environmental change, emphasizing the importance of local fish–habitat relationships in streams.

  8. A Statistical Method to Predict Flow Permanence in Dryland Streams from Time Series of Stream Temperature

    Directory of Open Access Journals (Sweden)

    Ivan Arismendi

    2017-12-01

    Full Text Available Intermittent and ephemeral streams represent more than half of the length of the global river network. Dryland freshwater ecosystems are especially vulnerable to changes in human-related water uses as well as shifts in terrestrial climates. Yet, the description and quantification of patterns of flow permanence in these systems is challenging mostly due to difficulties in instrumentation. Here, we took advantage of existing stream temperature datasets in dryland streams in the northwest Great Basin desert, USA, to extract critical information on climate-sensitive patterns of flow permanence. We used a signal detection technique, Hidden Markov Models (HMMs, to extract information from daily time series of stream temperature to diagnose patterns of stream drying. Specifically, we applied HMMs to time series of daily standard deviation (SD of stream temperature (i.e., dry stream channels typically display highly variable daily temperature records compared to wet stream channels between April and August (2015–2016. We used information from paired stream and air temperature data loggers as well as co-located stream temperature data loggers with electrical resistors as confirmatory sources of the timing of stream drying. We expanded our approach to an entire stream network to illustrate the utility of the method to detect patterns of flow permanence over a broader spatial extent. We successfully identified and separated signals characteristic of wet and dry stream conditions and their shifts over time. Most of our study sites within the entire stream network exhibited a single state over the entire season (80%, but a portion of them showed one or more shifts among states (17%. We provide recommendations to use this approach based on a series of simple steps. Our findings illustrate a successful method that can be used to rigorously quantify flow permanence regimes in streams using existing records of stream temperature.

  9. A statistical method to predict flow permanence in dryland streams from time series of stream temperature

    Science.gov (United States)

    Arismendi, Ivan; Dunham, Jason B.; Heck, Michael; Schultz, Luke; Hockman-Wert, David

    2017-01-01

    Intermittent and ephemeral streams represent more than half of the length of the global river network. Dryland freshwater ecosystems are especially vulnerable to changes in human-related water uses as well as shifts in terrestrial climates. Yet, the description and quantification of patterns of flow permanence in these systems is challenging mostly due to difficulties in instrumentation. Here, we took advantage of existing stream temperature datasets in dryland streams in the northwest Great Basin desert, USA, to extract critical information on climate-sensitive patterns of flow permanence. We used a signal detection technique, Hidden Markov Models (HMMs), to extract information from daily time series of stream temperature to diagnose patterns of stream drying. Specifically, we applied HMMs to time series of daily standard deviation (SD) of stream temperature (i.e., dry stream channels typically display highly variable daily temperature records compared to wet stream channels) between April and August (2015–2016). We used information from paired stream and air temperature data loggers as well as co-located stream temperature data loggers with electrical resistors as confirmatory sources of the timing of stream drying. We expanded our approach to an entire stream network to illustrate the utility of the method to detect patterns of flow permanence over a broader spatial extent. We successfully identified and separated signals characteristic of wet and dry stream conditions and their shifts over time. Most of our study sites within the entire stream network exhibited a single state over the entire season (80%), but a portion of them showed one or more shifts among states (17%). We provide recommendations to use this approach based on a series of simple steps. Our findings illustrate a successful method that can be used to rigorously quantify flow permanence regimes in streams using existing records of stream temperature.

  10. Relationship between bifenthrin sediment toxic units and benthic community metrics in urban California streams.

    Science.gov (United States)

    Hall, Lenwood W; Anderson, Ronald D

    2013-08-01

    The objective of this study was to use ecologically relevant field measurements for determining the relationship between bifenthrin sediment toxic units (TUs) (environmental concentrations/Hyalella acute LC50 value) and 15 benthic metrics in four urban California streams sampled from 2006 to 2011. Data from the following four California streams were used in the analysis: Kirker Creek (2006, 2007), Pleasant Grove Creek (2006, 2007, and 2008), Arcade Creek (2009, 2010, and 2011), and Salinas streams (2009, 2010, and 2011). The results from univariate analysis of benthic metrics versus bifenthrin TU calculations for the four California streams with multiple-year datasets combined by stream showed that there were either nonsignificant relationships or lack of metric data for 93 % of cases. For 7 % of the data (4 cases) where significant relationships were reported between benthic metrics and bifenthrin TUs, these relationships were ecologically meaningful. Three of these significant direct relationships were an expression of tolerant benthic taxa (either % tolerant taxa or tolerance values, which are similar metrics), which would be expected to increase in a stressed environment. These direct significant tolerance relationships were reported for Kirker Creek, Pleasant Grove Creek, and Arcade Creek. The fourth significant relationship was an inverse relationship between taxa richness and bifenthrin TUs for the 3-year Pleasant Grove Creek dataset. In summary, only a small percent of the benthic metric × bifenthrin TU relationships were significant for the four California streams. Therefore, the general summary conclusion from this analysis is that there is no strong case for showing consistent meaningful relationships between various benthic metrics used to characterize the status of benthic communities and bifenthrin TUs for these four California streams.

  11. Recurrent and Dynamic Models for Predicting Streaming Video Quality of Experience.

    Science.gov (United States)

    Bampis, Christos G; Li, Zhi; Katsavounidis, Ioannis; Bovik, Alan C

    2018-07-01

    Streaming video services represent a very large fraction of global bandwidth consumption. Due to the exploding demands of mobile video streaming services, coupled with limited bandwidth availability, video streams are often transmitted through unreliable, low-bandwidth networks. This unavoidably leads to two types of major streaming-related impairments: compression artifacts and/or rebuffering events. In streaming video applications, the end-user is a human observer; hence being able to predict the subjective Quality of Experience (QoE) associated with streamed videos could lead to the creation of perceptually optimized resource allocation strategies driving higher quality video streaming services. We propose a variety of recurrent dynamic neural networks that conduct continuous-time subjective QoE prediction. By formulating the problem as one of time-series forecasting, we train a variety of recurrent neural networks and non-linear autoregressive models to predict QoE using several recently developed subjective QoE databases. These models combine multiple, diverse neural network inputs, such as predicted video quality scores, rebuffering measurements, and data related to memory and its effects on human behavioral responses, using them to predict QoE on video streams impaired by both compression artifacts and rebuffering events. Instead of finding a single time-series prediction model, we propose and evaluate ways of aggregating different models into a forecasting ensemble that delivers improved results with reduced forecasting variance. We also deploy appropriate new evaluation metrics for comparing time-series predictions in streaming applications. Our experimental results demonstrate improved prediction performance that approaches human performance. An implementation of this work can be found at https://github.com/christosbampis/NARX_QoE_release.

  12. Numeric Analysis for Relationship-Aware Scalable Streaming Scheme

    Directory of Open Access Journals (Sweden)

    Heung Ki Lee

    2014-01-01

    Full Text Available Frequent packet loss of media data is a critical problem that degrades the quality of streaming services over mobile networks. Packet loss invalidates frames containing lost packets and other related frames at the same time. Indirect loss caused by losing packets decreases the quality of streaming. A scalable streaming service can decrease the amount of dropped multimedia resulting from a single packet loss. Content providers typically divide one large media stream into several layers through a scalable streaming service and then provide each scalable layer to the user depending on the mobile network. Also, a scalable streaming service makes it possible to decode partial multimedia data depending on the relationship between frames and layers. Therefore, a scalable streaming service provides a way to decrease the wasted multimedia data when one packet is lost. However, the hierarchical structure between frames and layers of scalable streams determines the service quality of the scalable streaming service. Even if whole packets of layers are transmitted successfully, they cannot be decoded as a result of the absence of reference frames and layers. Therefore, the complicated relationship between frames and layers in a scalable stream increases the volume of abandoned layers. For providing a high-quality scalable streaming service, we choose a proper relationship between scalable layers as well as the amount of transmitted multimedia data depending on the network situation. We prove that a simple scalable scheme outperforms a complicated scheme in an error-prone network. We suggest an adaptive set-top box (AdaptiveSTB to lower the dependency between scalable layers in a scalable stream. Also, we provide a numerical model to obtain the indirect loss of multimedia data and apply it to various multimedia streams. Our AdaptiveSTB enhances the quality of a scalable streaming service by removing indirect loss.

  13. Standards for Multi-Stream and Multi-Device Media Synchronisation

    NARCIS (Netherlands)

    Deventer, M.O. van; Stokking, H.M.; Hammond, M.; Cesar, P.

    2016-01-01

    Media synchronization is getting renewed attention with ecosystems of connected devices enabling novel media consumption paradigms. Social TV, hybrid TV, and companion screens are examples that are enabling people to consume multiple media streams at multiple devices together. These novel use cases

  14. Extending Data Worth Analyses to Select Multiple Observations Targeting Multiple Forecasts.

    Science.gov (United States)

    Vilhelmsen, Troels N; Ferré, Ty P A

    2017-09-15

    Hydrological models are often set up to provide specific forecasts of interest. Owing to the inherent uncertainty in data used to derive model structure and used to constrain parameter variations, the model forecasts will be uncertain. Additional data collection is often performed to minimize this forecast uncertainty. Given our common financial restrictions, it is critical that we identify data with maximal information content with respect to forecast of interest. In practice, this often devolves to qualitative decisions based on expert opinion. However, there is no assurance that this will lead to optimal design, especially for complex hydrogeological problems. Specifically, these complexities include considerations of multiple forecasts, shared information among potential observations, information content of existing data, and the assumptions and simplifications underlying model construction. In the present study, we extend previous data worth analyses to include: simultaneous selection of multiple new measurements and consideration of multiple forecasts of interest. We show how the suggested approach can be used to optimize data collection. This can be used in a manner that suggests specific measurement sets or that produces probability maps indicating areas likely to be informative for specific forecasts. Moreover, we provide examples documenting that sequential measurement election approaches often lead to suboptimal designs and that estimates of data covariance should be included when selecting future measurement sets. © 2017, National Ground Water Association.

  15. Yosemite Hydroclimate Network: Distributed stream and atmospheric data for the Tuolumne River watershed and surroundings

    Science.gov (United States)

    Lundquist, Jessica D.; Roche, James W.; Forrester, Harrison; Moore, Courtney; Keenan, Eric; Perry, Gwyneth; Cristea, Nicoleta; Henn, Brian; Lapo, Karl; McGurk, Bruce; Cayan, Daniel R.; Dettinger, Michael D.

    2016-01-01

    Regions of complex topography and remote wilderness terrain have spatially varying patterns of temperature and streamflow, but due to inherent difficulties of access, are often very poorly sampled. Here we present a data set of distributed stream stage, streamflow, stream temperature, barometric pressure, and air temperature from the Tuolumne River Watershed in Yosemite National Park, Sierra Nevada, California, USA, for water years 2002–2015, as well as a quality-controlled hourly meteorological forcing time series for use in hydrologic modeling. We also provide snow data and daily inflow to the Hetch Hetchy Reservoir for 1970–2015. This paper describes data collected using low-visibility and low-impact installations for wilderness locations and can be used alone or as a critical supplement to ancillary data sets collected by cooperating agencies, referenced herein. This data set provides a unique opportunity to understand spatial patterns and scaling of hydroclimatic processes in complex terrain and can be used to evaluate downscaling techniques or distributed modeling. The paper also provides an example methodology and lessons learned in conducting hydroclimatic monitoring in remote wilderness.

  16. How wide is a stream? Spatial extent of the potential "stream signature" in terrestrial food webs using meta-analysis.

    Science.gov (United States)

    Muehlbauer, Jeffrey D; Collins, Scott F; Doyle, Martin W; Tockner, Klement

    2014-01-01

    The magnitude of cross-ecosystem resource subsidies is increasingly well recognized; however, less is known about the distance these subsidies travel into the recipient landscape. In streams and rivers, this distance can delimit the "biological stream width," complementary to hydro-geomorphic measures (e.g., channel banks) that have typically defined stream ecosystem boundaries. In this study we used meta-analysis to define a "stream signature" on land that relates the stream-to-land subsidy to distance. The 50% stream signature, for example, identifies the point on the landscape where subsidy resources are still at half of their maximum (in- or near-stream) level. The decay curve for these data was best fit by a negative power function in which the 50% stream signature was concentrated near stream banks (1.5 m), but a non-trivial (10%) portion of the maximum subsidy level was still found > 0.5 km from the water's edge. The meta-analysis also identified explanatory variables that affect the stream signature. This improves our understanding of ecosystem conditions that permit spatially extensive subsidy transmission, such as in highly productive, middle-order streams and rivers. Resultant multivariate models from this analysis may be useful to managers implementing buffer rules and conservation strategies for stream and riparian function, as they facilitate prediction of the extent of subsidies. Our results stress that much of the subsidy remains near the stream, but also that subsidies (and aquatic organisms) are capable of long-distance dispersal into adjacent environments, and that the effective "biological stream width" of stream and river ecosystems is often much larger than has been defined by hydro-geomorphic metrics alone. Limited data available from marine and lake sources overlap well with the stream signature data, indicating that the "signature" approach may also be applicable to subsidy spatial dynamics across other ecosystems.

  17. Modeling and clustering users with evolving profiles in usage streams

    KAUST Repository

    Zhang, Chongsheng; Masseglia, Florent; Zhang, Xiangliang

    2012-01-01

    Today, there is an increasing need of data stream mining technology to discover important patterns on the fly. Existing data stream models and algorithms commonly assume that users' records or profiles in data streams will not be updated or revised once they arrive. Nevertheless, in various applications such asWeb usage, the records/profiles of the users can evolve along time. This kind of streaming data evolves in two forms, the streaming of tuples or transactions as in the case of traditional data streams, and more importantly, the evolving of user records/profiles inside the streams. Such data streams bring difficulties on modeling and clustering for exploring users' behaviors. In this paper, we propose three models to summarize this kind of data streams, which are the batch model, the Evolving Objects (EO) model and the Dynamic Data Stream (DDS) model. Through creating, updating and deleting user profiles, these models summarize the behaviors of each user as a profile object. Based upon these models, clustering algorithms are employed to discover interesting user groups from the profile objects. We have evaluated all the proposed models on a large real-world data set, showing that the DDS model summarizes the data streams with evolving tuples more efficiently and effectively, and provides better basis for clustering users than the other two models. © 2012 IEEE.

  18. Modeling and clustering users with evolving profiles in usage streams

    KAUST Repository

    Zhang, Chongsheng

    2012-09-01

    Today, there is an increasing need of data stream mining technology to discover important patterns on the fly. Existing data stream models and algorithms commonly assume that users\\' records or profiles in data streams will not be updated or revised once they arrive. Nevertheless, in various applications such asWeb usage, the records/profiles of the users can evolve along time. This kind of streaming data evolves in two forms, the streaming of tuples or transactions as in the case of traditional data streams, and more importantly, the evolving of user records/profiles inside the streams. Such data streams bring difficulties on modeling and clustering for exploring users\\' behaviors. In this paper, we propose three models to summarize this kind of data streams, which are the batch model, the Evolving Objects (EO) model and the Dynamic Data Stream (DDS) model. Through creating, updating and deleting user profiles, these models summarize the behaviors of each user as a profile object. Based upon these models, clustering algorithms are employed to discover interesting user groups from the profile objects. We have evaluated all the proposed models on a large real-world data set, showing that the DDS model summarizes the data streams with evolving tuples more efficiently and effectively, and provides better basis for clustering users than the other two models. © 2012 IEEE.

  19. Population persistence of stream fish in response to environmental change: integrating data and models across space

    Science.gov (United States)

    Letcher, B. H.; Schueller, P.; Bassar, R.; Coombs, J.; Rosner, A.; Sakrejda, K.; Kanno, Y.; Whiteley, A.; Nislow, K. H.

    2013-12-01

    For stream fishes, environmental variation is a key driver of individual body growth/movement/survival and, by extension, population dynamics. Identifying how stream fish respond to environmental variation can help clarify mechanisms responsible for population dynamics and can help provide tools to forecast relative resilience of populations across space. Forecasting dynamics across space is challenging, however, because it can be difficult to conduct enough studies with enough intensity to fully characterize broad-scale population response to environmental change. We have adopted a multi-scale approach, using detailed individual-based studies and analyses (integral projection matrix) to determine sensitivities of population growth to environmental variation combined with broad spatial data and analyses (occupancy and abundance models) to estimate patterns of population response across space. Population growth of brook trout was most sensitive to stream flow in the spring and winter, most sensitive to stream temperature in the fall and sensitive to both flow and temperature in the summer. High flow in the spring and winter had negative effects on population growth while high temperature had a negative effect in the fall. Flow had no effect when it was cold, but a positive effect when it was warm in the summer. Combined with occupancy and abundance models, these data give insight into the spatial structure of resilient populations and can help guide prioritization of management actions.

  20. Hydrogeochemical and stream sediment reconnaissance basic data for Dallas NTMS Quadrangle, Texas

    International Nuclear Information System (INIS)

    1981-01-01

    Results of a reconnaissance geochemical survey of the Dallas Quadrangle, Texas are reported. Field and laboratory data are presented for 284 groundwater and 545 stream sediment samples. Statistical and areal distribution plots of uranium and possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided. Groundwater produced from the Navarro Group, Neylandville Formation, Marlbrook Marl, and the Glen Rose and Twin Mountains Formations exhibit anomalous uranium (> 9.05 ppB) and specific conductance (> 1871 μmhos/cm) values. The anomalies represent a southern extension of a similar trend observed in the Sherman Quadrangle, K/UR-110. Stream sediments representing the Eagle Ford Group and Woodbine Formation exhibit the highest concentrations of total and hot-acid-soluble uranium and thorium of samples collected in the Dallas Quadrangle. The U/TU value indicates that > 80% of this uranium is present in a soluble form

  1. Computer analysis to the geochemical interpretation of soil and stream sediment data in an area of Southern Uruguay

    International Nuclear Information System (INIS)

    Spangenberg, J.

    2010-01-01

    In southern Uruguay there are several known occurrences of base metal sulphide mineralization within an area of Precambrian volcanic sedimentary rocks. Regional geochemical stream sediment reconnaissance surveys revealed new polymetallic anomalies in the same stratigraphic zone. Geochemical interpretation of multi-element data from a soil and stream sediment survey carried out in one of these anomalous areas is presented.

  2. Multiple streams approach to tobacco control policymaking in a tobacco-growing state.

    Science.gov (United States)

    Mamudu, Hadii M; Dadkar, Sumati; Veeranki, Sreenivas P; He, Yi; Barnes, Richard; Glantz, Stanton A

    2014-08-01

    Smokefree policies (SFPs) have diffused throughout the US and worldwide. However, the development of SFPs in the difficult policy environment of tobacco-producing states and economies worldwide has not been well-explored. In 2007, Tennessee, the third largest tobacco producer in the US, enacted the Non-Smoker Protection Act (NSPA). This study utilizes the multiple streams model to provide understanding of why and how this policy was developed by triangulating interviews with key stakeholders and legislative debates with archival documents. In June 2006, the Governor unexpectedly announced support for SFP, which created a window of opportunity for policy change. The Campaign for Healthy and Responsible Tennessee, a health coalition, seized this opportunity and worked with the administration and the Tennessee Restaurant Association to negotiate a comprehensive SFP, however, a weaker bill was used by the legislative leadership to develop the NSPA. Although the Governor and the Tennessee Restaurant Association's support generated an environment for 100% SFP, health groups did not fully capitalize on this environmental change and settled for a weak policy with several exemptions. This study suggests the importance for proponents of policy change to understand changes in their environment and be willing and able to capitalize on these changes.

  3. Real-Time Management of Multimodal Streaming Data for Monitoring of Epileptic Patients.

    Science.gov (United States)

    Triantafyllopoulos, Dimitrios; Korvesis, Panagiotis; Mporas, Iosif; Megalooikonomou, Vasileios

    2016-03-01

    New generation of healthcare is represented by wearable health monitoring systems, which provide real-time monitoring of patient's physiological parameters. It is expected that continuous ambulatory monitoring of vital signals will improve treatment of patients and enable proactive personal health management. In this paper, we present the implementation of a multimodal real-time system for epilepsy management. The proposed methodology is based on a data streaming architecture and efficient management of a big flow of physiological parameters. The performance of this architecture is examined for varying spatial resolution of the recorded data.

  4. HUNTING THE PARENT OF THE ORPHAN STREAM: IDENTIFYING STREAM MEMBERS FROM LOW-RESOLUTION SPECTROSCOPY

    International Nuclear Information System (INIS)

    Casey, Andrew R.; Da Costa, Gary; Keller, Stefan C.; Maunder, Elizabeth

    2013-01-01

    We present candidate K-giant members in the Orphan Stream that have been identified from low-resolution data taken with the AAOmega spectrograph on the Anglo-Australian Telescope. From modest signal-to-noise spectra and independent cuts in photometry, kinematics, gravity, and metallicity we yield self-consistent, highly probable stream members. We find a revised stream distance of 22.5 ± 2.0 kpc near the celestial equator and our kinematic signature peaks at V GSR = 82.1 ± 1.4 km s –1 . The observed velocity dispersion of our most probable members is consistent with arising from the velocity uncertainties alone. This indicates that at least along this line of sight, the Orphan Stream is kinematically cold. Our data indicate an overall stream metallicity of [Fe/H] = –1.63 ± 0.19 dex which is more metal-rich than previously found and unbiased by spectral type. Furthermore, the significant metallicity dispersion displayed by our most probable members, σ([Fe/H]) = 0.56 dex, suggests that the unidentified Orphan Stream parent is a dSph satellite. We highlight likely members for high-resolution spectroscopic follow-up.

  5. Analytical data and sample locality map for aqua-regia leachates of stream sediments analyzed by ICP, and emission spectrographic and ICP results for many NURE stream sediments from the Killik River Quadrangle, Alaska

    International Nuclear Information System (INIS)

    Motooka, J.M.; Adrian, B.M.; Church, S.E.; McDougal, C.M.; Fife, J.B.

    1989-01-01

    A U.S. Geological Survey report is presented giving analytical data and sample locality map for aqua-regia leachates of stream sediments analyzed by ICP, and emission spectrographic and ICP results for many NURE stream sediments from the Killik River Quadrangle, Alaska

  6. Death Valley 10 x 20 NTMS area, California and Nevada. Data report: National Uranium Resource Evaluation program, hydrogeochemical and stream sediment reconnaissance

    International Nuclear Information System (INIS)

    Cook, J.R.

    1980-04-01

    Results of ground water and stream sediment reconnaissance in the National Topographic Map Series (NTMS) Death Valley 1 0 x 2 0 quadrangle are presented. Stream sediment samples were collected from small streams at 649 sites or at a nominal density of one site per 20 square kilometers. Ground water samples were collected at 62 sites or at a nominal density of one site per 220 square kilometers. Neutron activation analysis results are given for uranium and 16 other elements in sediments, and for uranium and 8 other elements in ground water and surface water. Mass spectrometry results are given for helium in ground water. Field measurements and observations are reported for each site. Analytical data and field measurements are presented in tables and maps. Statistical summaries of data and a brief description of results are given. A generalized geologic map and a summary of the geology of the area are included. Key data from ground water sites include (1) water chemistry measurements (pH, conductivity, and alkalinity), (2) scintillometer readings, and (3) elemental analyses (U, Br, Cl, F, He, Mn, Na, and V). Supplementary data include site descriptors, tabulated analytical data for Al, Dy, and Mg, and histograms and cumulative frequency plots for all elements. Key data from stream sediment sites include (1) water quality measurements (2) important elemental analyses, (U, Th, Hf, Ce, Fe, Mn, Sc, Na, Ti, and V), and (3) scintillometer readings. Supplementary data from stream sediment sites include sample site descriptors (stream characteristics, vegetation, etc.), additional elemental analyses (Dy, Eu, La, Lu, Sm, and Yb), and histograms and cumulative frequency plots for all elements

  7. ThermoData Engine: Extension to Solvent Design and Multi-component Process Stream Property Calculations with Uncertainty Analysis

    DEFF Research Database (Denmark)

    Diky, Vladimir; Chirico, Robert D.; Muzny, Chris

    ThermoData Engine (TDE, NIST Standard Reference Databases 103a and 103b) is the first product that implements the concept of Dynamic Data Evaluation in the fields of thermophysics and thermochemistry, which includes maintaining the comprehensive and up-to-date database of experimentally measured...... property values and expert system for data analysis and generation of recommended property values at the specified conditions along with uncertainties on demand. The most recent extension of TDE covers solvent design and multi-component process stream property calculations with uncertainty analysis...... variations). Predictions can be compared to the available experimental data, and uncertainties are estimated for all efficiency criteria. Calculations of the properties of multi-component streams including composition at phase equilibria (flash calculations) are at the heart of process simulation engines...

  8. Can Low-Resolution Airborne Laser Scanning Data Be Used to Model Stream Rating Curves?

    Directory of Open Access Journals (Sweden)

    Steve W. Lyon

    2015-03-01

    Full Text Available This pilot study explores the potential of using low-resolution (0.2 points/m2 airborne laser scanning (ALS-derived elevation data to model stream rating curves. Rating curves, which allow the functional translation of stream water depth into discharge, making them integral to water resource monitoring efforts, were modeled using a physics-based approach that captures basic geometric measurements to establish flow resistance due to implicit channel roughness. We tested synthetically thinned high-resolution (more than 2 points/m2 ALS data as a proxy for low-resolution data at a point density equivalent to that obtained within most national-scale ALS strategies. Our results show that the errors incurred due to the effect of low-resolution versus high-resolution ALS data were less than those due to flow measurement and empirical rating curve fitting uncertainties. As such, although there likely are scale and technical limitations to consider, it is theoretically possible to generate rating curves in a river network from ALS data of the resolution anticipated within national-scale ALS schemes (at least for rivers with relatively simple geometries. This is promising, since generating rating curves from ALS scans would greatly enhance our ability to monitor streamflow by simplifying the overall effort required.

  9. Can low-resolution airborne laser scanning data be used to model stream rating curves?

    Science.gov (United States)

    Lyon, Steve; Nathanson, Marcus; Lam, Norris; Dahlke, Helen; Rutzinger, Martin; Kean, Jason W.; Laudon, Hjalmar

    2015-01-01

    This pilot study explores the potential of using low-resolution (0.2 points/m2) airborne laser scanning (ALS)-derived elevation data to model stream rating curves. Rating curves, which allow the functional translation of stream water depth into discharge, making them integral to water resource monitoring efforts, were modeled using a physics-based approach that captures basic geometric measurements to establish flow resistance due to implicit channel roughness. We tested synthetically thinned high-resolution (more than 2 points/m2) ALS data as a proxy for low-resolution data at a point density equivalent to that obtained within most national-scale ALS strategies. Our results show that the errors incurred due to the effect of low-resolution versus high-resolution ALS data were less than those due to flow measurement and empirical rating curve fitting uncertainties. As such, although there likely are scale and technical limitations to consider, it is theoretically possible to generate rating curves in a river network from ALS data of the resolution anticipated within national-scale ALS schemes (at least for rivers with relatively simple geometries). This is promising, since generating rating curves from ALS scans would greatly enhance our ability to monitor streamflow by simplifying the overall effort required.

  10. Streams with Strahler Stream Order

    Data.gov (United States)

    Minnesota Department of Natural Resources — Stream segments with Strahler stream order values assigned. As of 01/08/08 the linework is from the DNR24K stream coverages and will not match the updated...

  11. Contaminant Dynamics and Trends in Hyperalkaline Urban Streams

    Science.gov (United States)

    Riley, Alex; Mayes, William

    2015-04-01

    Streams in post-industrial urban areas can have multiple contemporary and historic pressures impacting upon their chemical and ecological status. This paper presents analysis of long term data series (up to 36 years in length) from two small streams in northern England (catchment areas 0.5-0.6km2). Around 3.5 million m3 of steel making slags and other wastes were deposited in the headwater areas of the Howden Burn and Dene Burn in northeast England up to the closure of the workings in the early 1980s. This has led to streams draining from the former workings which have a hyperalkaline ambient pH (mean of 10.3 in both streams), elevated alkalinity (up to 487 mg/L as CaCO3) from leaching of lime and other calcium oxides / silicates within the slag, and enrichment of some trace elements (e.g. aluminium (Al), lithium (Li) and zinc (Zn)) including those which form oxyanions mobile at high pH such as vanadium (V). The high ionic strength of the waters and calcium enrichment also leads to waters highly supersaturated with calcium carbonate. Trace contaminant concentrations are strongly positively correlated, and concentrations generally diminish with increased flow rate suggesting the key source of metals in the system is the highly alkaline groundwater draining from the slag mounds. Some contaminants (notably Cr and ammonium) increase with high flow suggesting sources related to urban runoff and drainage from combined sewer overflows into one of the catchments. Loading estimates instream show that many of the contaminants (e.g. Al, V and Zn) are rapidly attenuated in secondary calcium carbonate-dominated deposits that precipitate vigorously on the streambeds with rates of up to 250 g CaCO3/m2/day. These secondary sinks limit the mobility of many contaminants in the water column, while concentrations in secondary deposits are relatively low given the rapid rates at which Ca is also attenuated. Long-term trend analysis showed modest declines in calcium and alkalinity over

  12. Network Transfer of Control Data: An Application of the NIST SMART DATA FLOW

    Directory of Open Access Journals (Sweden)

    Vincent Stanford

    2004-12-01

    Full Text Available Pervasive Computing environments range from basic mobile point of sale terminal systems, to rich Smart Spaces with many devices and sensors such as lapel microphones, audio and video sensor arrays and multiple interactive PDA acting as electronic brief cases, providing authentication, and user preference data to the environment. These systems present new challenges in distributed human-computer interfaces such as how to best use sensor streams, distribute interfaces across multiple devices, and dynamic network management as users come an go, and as devices are added or fail. The NATIONAL INSTITUTE OF STANDARDS AND TECHNOLOGY SMART DATA FLOW system is a low overhead, high bandwidth transport mechanism for standardized multi-modal data streams. It is designed to allow integration of multiple sensors with distributed processing needed for the sense-recognize-respond cycle of multi modal user interfaces. Its core is a server/client architecture, allowing clients to produce or subscribe to data flows, and supporting steps toward scalable processing, distributing the computing requirements among many network connected computers and pervasive devices. This article introduces the communication broker and provides an example of an effective real time sensor fusion to track a speaker with a video camera using data captured from multi-channel microphone array.

  13. The kinematic footprints of five stellar streams in Andromeda's halo

    Science.gov (United States)

    Chapman, S. C.; Ibata, R.; Irwin, M.; Koch, A.; Letarte, B.; Martin, N.; Collins, M.; Lewis, G. F.; McConnachie, A.; Peñarrubia, J.; Rich, R. M.; Trethewey, D.; Ferguson, A.; Huxor, A.; Tanvir, N.

    2008-11-01

    halo is largely made up of multiple kinematically cold streams. The data presented herein were obtained at the W. M. Keck Observatory, which is operated as a scientific partnership among the California Institute of Technology, the University of California and the National Aeronautics and Space Administration. The Observatory was made possible by the generous financial support of the W. M. Keck Foundation. E-mail: schapman@ast.cam.ac.uk ‡ Canadian Space Agency, Space Science Fellow.

  14. Autonomous Byte Stream Randomizer

    Science.gov (United States)

    Paloulian, George K.; Woo, Simon S.; Chow, Edward T.

    2013-01-01

    Net-centric networking environments are often faced with limited resources and must utilize bandwidth as efficiently as possible. In networking environments that span wide areas, the data transmission has to be efficient without any redundant or exuberant metadata. The Autonomous Byte Stream Randomizer software provides an extra level of security on top of existing data encryption methods. Randomizing the data s byte stream adds an extra layer to existing data protection methods, thus making it harder for an attacker to decrypt protected data. Based on a generated crypto-graphically secure random seed, a random sequence of numbers is used to intelligently and efficiently swap the organization of bytes in data using the unbiased and memory-efficient in-place Fisher-Yates shuffle method. Swapping bytes and reorganizing the crucial structure of the byte data renders the data file unreadable and leaves the data in a deconstructed state. This deconstruction adds an extra level of security requiring the byte stream to be reconstructed with the random seed in order to be readable. Once the data byte stream has been randomized, the software enables the data to be distributed to N nodes in an environment. Each piece of the data in randomized and distributed form is a separate entity unreadable on its own right, but when combined with all N pieces, is able to be reconstructed back to one. Reconstruction requires possession of the key used for randomizing the bytes, leading to the generation of the same cryptographically secure random sequence of numbers used to randomize the data. This software is a cornerstone capability possessing the ability to generate the same cryptographically secure sequence on different machines and time intervals, thus allowing this software to be used more heavily in net-centric environments where data transfer bandwidth is limited.

  15. Combining Empirical Relationships with Data Based Mechanistic Modeling to Inform Solute Tracer Investigations across Stream Orders

    Science.gov (United States)

    Herrington, C.; Gonzalez-Pinzon, R.; Covino, T. P.; Mortensen, J.

    2015-12-01

    Solute transport studies in streams and rivers often begin with the introduction of conservative and reactive tracers into the water column. Information on the transport of these substances is then captured within tracer breakthrough curves (BTCs) and used to estimate, for instance, travel times and dissolved nutrient and carbon dynamics. Traditionally, these investigations have been limited to systems with small discharges (turbidity (e.g., nitrate signals with SUNA instruments or fluorescence measures) and/or high total dissolved solids (e.g., making prohibitively expensive the use of salt tracers such as NaCl) in larger systems. Additionally, a successful time-of-travel study is valuable for only a single discharge and river stage. We have developed a method to predict tracer BTCs to inform sampling frequencies at small and large stream orders using empirical relationships developed from multiple tracer injections spanning several orders of magnitude in discharge and reach length. This method was successfully tested in 1st to 8th order systems along the Middle Rio Grande River Basin in New Mexico, USA.

  16. Stream Habitat Reach Summary - NCWAP [ds158

    Data.gov (United States)

    California Natural Resource Agency — The Stream Habitat - NCWAP - Reach Summary [ds158] shapefile contains in-stream habitat survey data summarized to the stream reach level. It is a derivative of the...

  17. DEVELOPMENT OF TRANSPORT SUBSYSTEM STREAMING DATA REPLICATION CLUSTER IN CORBA-SYSTEM WITH ZEROMQ TECHNOLOGY

    Directory of Open Access Journals (Sweden)

    F. A. Kozlov

    2013-03-01

    Full Text Available The article deals with the peculiarities of distributed cluster system creation with streaming data replication. Ways of replication cluster implementation in CORBA-systems with ZeroMq technology are presented. Major advantages of ZeroMQ technology over similar technologies are considered in this type distributed systems creation.

  18. Estimating discharge and non-point source nitrate loading to streams from three end-member pathways using high-frequency water quality and streamflow data

    Science.gov (United States)

    Miller, M. P.; Tesoriero, A. J.; Hood, K.; Terziotti, S.; Wolock, D.

    2017-12-01

    The myriad hydrologic and biogeochemical processes taking place in watersheds occurring across space and time are integrated and reflected in the quantity and quality of water in streams and rivers. Collection of high-frequency water quality data with sensors in surface waters provides new opportunities to disentangle these processes and quantify sources and transport of water and solutes in the coupled groundwater-surface water system. A new approach for separating the streamflow hydrograph into three components was developed and coupled with high-frequency specific conductance and nitrate data to estimate time-variable watershed-scale nitrate loading from three end-member pathways - dilute quickflow, concentrated quickflow, and slowflow groundwater - to two streams in central Wisconsin. Time-variable nitrate loads from the three pathways were estimated for periods of up to two years in a groundwater-dominated and a quickflow-dominated stream, using only streamflow and in-stream water quality data. The dilute and concentrated quickflow end-members were distinguished using high-frequency specific conductance data. Results indicate that dilute quickflow contributed less than 5% of the nitrate load at both sites, whereas 89±5% of the nitrate load at the groundwater-dominated stream was from slowflow groundwater, and 84±13% of the nitrate load at the quickflow-dominated stream was from concentrated quickflow. Concentrated quickflow nitrate concentrations varied seasonally at both sites, with peak concentrations in the winter that were 2-3 times greater than minimum concentrations during the growing season. Application of this approach provides an opportunity to assess stream vulnerability to non-point source nitrate loading and expected stream responses to current or changing conditions and practices in watersheds.

  19. Multiple variables data sets visualization in ROOT

    International Nuclear Information System (INIS)

    Couet, O

    2008-01-01

    The ROOT graphical framework provides support for many different functions including basic graphics, high-level visualization techniques, output on files, 3D viewing etc. They use well-known world standards to render graphics on screen, to produce high-quality output files, and to generate images for Web publishing. Many techniques allow visualization of all the basic ROOT data types, but the graphical framework was still a bit weak in the visualization of multiple variables data sets. This paper presents latest developments done in the ROOT framework to visualize multiple variables (>4) data sets

  20. Relative Linkages of Stream Dissolved Oxygen with the Hydroclimatic and Biogeochemical Drivers across the Gulf Coast of U.S.A.

    Science.gov (United States)

    Gebreslase, A. K.; Abdul-Aziz, O. I.

    2017-12-01

    Dynamics of coastal stream water quality is influenced by a multitude of interacting environmental drivers. A systematic data analytics approach was employed to determine the relative linkages of stream dissolved oxygen (DO) with the hydroclimatic and biogeochemical variables across the Gulf Coast of U.S.A. Multivariate pattern recognition techniques of PCA and FA, alongside Pearson's correlation matrix, were utilized to examine the interrelation of variables at 36 water quality monitoring stations from USGS NWIS and EPA STORET databases. Power-law based partial least square regression models with a bootstrap Monte Carlo procedure (1000 iterations) were developed to estimate the relative linkages of dissolved oxygen with the hydroclimatic and biogeochemical variables by appropriately resolving multicollinearity (Nash-Sutcliffe efficiency = 0.58-0.94). Based on the dominant drivers, stations were divided into four environmental regimes. Water temperature was the dominant driver of DO in the majority of streams, representing most the northern part of Gulf Coast states. However, streams in the southern part of Texas and Florida showed a dominant pH control on stream DO. Further, streams representing the transition zone of the two environmental regimes showed notable controls of multiple drivers (i.e., water temperature, stream flow, and specific conductance) on the stream DO. The data analytics research provided profound insight to understand the dynamics of stream DO with the hydroclimatic and biogeochemical variables. The knowledge can help water quality managers in formulating plans for effective stream water quality and watershed management in the U.S. Gulf Coast. Keywords Data analytics, coastal streams, relative linkages, dissolved oxygen, environmental regimes, Gulf Coast, United States.

  1. Convergent Time-Varying Regression Models for Data Streams: Tracking Concept Drift by the Recursive Parzen-Based Generalized Regression Neural Networks.

    Science.gov (United States)

    Duda, Piotr; Jaworski, Maciej; Rutkowski, Leszek

    2018-03-01

    One of the greatest challenges in data mining is related to processing and analysis of massive data streams. Contrary to traditional static data mining problems, data streams require that each element is processed only once, the amount of allocated memory is constant and the models incorporate changes of investigated streams. A vast majority of available methods have been developed for data stream classification and only a few of them attempted to solve regression problems, using various heuristic approaches. In this paper, we develop mathematically justified regression models working in a time-varying environment. More specifically, we study incremental versions of generalized regression neural networks, called IGRNNs, and we prove their tracking properties - weak (in probability) and strong (with probability one) convergence assuming various concept drift scenarios. First, we present the IGRNNs, based on the Parzen kernels, for modeling stationary systems under nonstationary noise. Next, we extend our approach to modeling time-varying systems under nonstationary noise. We present several types of concept drifts to be handled by our approach in such a way that weak and strong convergence holds under certain conditions. Finally, in the series of simulations, we compare our method with commonly used heuristic approaches, based on forgetting mechanism or sliding windows, to deal with concept drift. Finally, we apply our concept in a real life scenario solving the problem of currency exchange rates prediction.

  2. Nutrient spiraling in streams and river networks

    Science.gov (United States)

    Ensign, Scott H.; Doyle, Martin W.

    2006-12-01

    Over the past 3 decades, nutrient spiraling has become a unifying paradigm for stream biogeochemical research. This paper presents (1) a quantitative synthesis of the nutrient spiraling literature and (2) application of these data to elucidate trends in nutrient spiraling within stream networks. Results are based on 404 individual experiments on ammonium (NH4), nitrate (NO3), and phosphate (PO4) from 52 published studies. Sixty-nine percent of the experiments were performed in first- and second-order streams, and 31% were performed in third- to fifth-order streams. Uptake lengths, Sw, of NH4 (median = 86 m) and PO4 (median = 96 m) were significantly different (α = 0.05) than NO3 (median = 236 m). Areal uptake rates of NH4 (median = 28 μg m-2 min-1) were significantly different than NO3 and PO4 (median = 15 and 14 μg m-2 min-1, respectively). There were significant differences among NH4, NO3, and PO4 uptake velocity (median = 5, 1, and 2 mm min-1, respectively). Correlation analysis results were equivocal on the effect of transient storage on nutrient spiraling. Application of these data to a stream network model showed that recycling (defined here as stream length ÷ Sw) of NH4 and NO3 generally increased with stream order, while PO4 recycling remained constant along a first- to fifth-order stream gradient. Within this hypothetical stream network, cumulative NH4 uptake decreased slightly with stream order, while cumulative NO3 and PO4 uptake increased with stream order. These data suggest the importance of larger rivers to nutrient spiraling and the need to consider how stream networks affect nutrient flux between terrestrial and marine ecosystems.

  3. Microtubule–microtubule sliding by kinesin-1 is essential for normal cytoplasmic streaming in Drosophila oocytes

    Science.gov (United States)

    Lu, Wen; Winding, Michael; Lakonishok, Margot; Wildonger, Jill

    2016-01-01

    Cytoplasmic streaming in Drosophila oocytes is a microtubule-based bulk cytoplasmic movement. Streaming efficiently circulates and localizes mRNAs and proteins deposited by the nurse cells across the oocyte. This movement is driven by kinesin-1, a major microtubule motor. Recently, we have shown that kinesin-1 heavy chain (KHC) can transport one microtubule on another microtubule, thus driving microtubule–microtubule sliding in multiple cell types. To study the role of microtubule sliding in oocyte cytoplasmic streaming, we used a Khc mutant that is deficient in microtubule sliding but able to transport a majority of cargoes. We demonstrated that streaming is reduced by genomic replacement of wild-type Khc with this sliding-deficient mutant. Streaming can be fully rescued by wild-type KHC and partially rescued by a chimeric motor that cannot move organelles but is active in microtubule sliding. Consistent with these data, we identified two populations of microtubules in fast-streaming oocytes: a network of stable microtubules anchored to the actin cortex and free cytoplasmic microtubules that moved in the ooplasm. We further demonstrated that the reduced streaming in sliding-deficient oocytes resulted in posterior determination defects. Together, we propose that kinesin-1 slides free cytoplasmic microtubules against cortically immobilized microtubules, generating forces that contribute to cytoplasmic streaming and are essential for the refinement of posterior determinants. PMID:27512034

  4. Stream fish colonization but not persistence varies regionally across a large North American river basin

    Science.gov (United States)

    Wheeler, Kit; Wengerd, Seth J.; Walsh, Stephen J.; Martin, Zachary P.; Jelks, Howard L.; Freeman, Mary C.

    2018-01-01

    Many species have distributions that span distinctly different physiographic regions, and effective conservation of such taxa will require a full accounting of all factors that potentially influence populations. Ecologists recognize effects of physiographic differences in topography, geology and climate on local habitat configurations, and thus the relevance of landscape heterogeneity to species distributions and abundances. However, research is lacking that examines how physiography affects the processes underlying metapopulation dynamics. We used data describing occupancy dynamics of stream fishes to evaluate evidence that physiography influences rates at which individual taxa persist in or colonize stream reaches under different flow conditions. Using periodic survey data from a stream fish assemblage in a large river basin that encompasses multiple physiographic regions, we fit multi-species dynamic occupancy models. Our modeling results suggested that stream fish colonization but not persistence was strongly governed by physiography, with estimated colonization rates considerably higher in Coastal Plain streams than in Piedmont and Blue Ridge systems. Like colonization, persistence was positively related to an index of stream flow magnitude, but the relationship between flow and persistence did not depend on physiography. Understanding the relative importance of colonization and persistence, and how one or both processes may change across the landscape, is critical information for the conservation of broadly distributed taxa, and conservation strategies explicitly accounting for spatial variation in these processes are likely to be more successful for such taxa.

  5. Fish populations in Plynlimon streams

    Directory of Open Access Journals (Sweden)

    D. T. Crisp

    1997-01-01

    Full Text Available In Plynlimon streams, brown trout (Salmo trutta L. are widespread in the upper Wye at population densities of 0.03 to 0.32 fish m-2 and show evidence of successful recruitment in most years. In the upper Severn, brown trout are found only in an area of c. 1670 -2 downstream of Blaenhafren Falls at densities of 0.03 to 0.24 fish -2 and the evidence suggests very variable year to year success in recruitment (Crisp & Beaumont, 1996. Analyses of the data show that temperature differences between afforested and unafforested streams may affect the rates of trout incubation and growth but are not likely to influence species survival. Simple analyses of stream discharge data suggest, but do not prove, that good years for recruitment in the Hafren population were years of low stream discharge. This may be linked to groundwater inputs detected in other studies in this stream. More research is needed to explain the survival of the apparently isolated trout population in the Hafren.

  6. Stream Deniable-Encryption Algorithms

    Directory of Open Access Journals (Sweden)

    N.A. Moldovyan

    2016-04-01

    Full Text Available A method for stream deniable encryption of secret message is proposed, which is computationally indistinguishable from the probabilistic encryption of some fake message. The method uses generation of two key streams with some secure block cipher. One of the key streams is generated depending on the secret key and the other one is generated depending on the fake key. The key streams are mixed with the secret and fake data streams so that the output ciphertext looks like the ciphertext produced by some probabilistic encryption algorithm applied to the fake message, while using the fake key. When the receiver or/and sender of the ciphertext are coerced to open the encryption key and the source message, they open the fake key and the fake message. To disclose their lie the coercer should demonstrate possibility of the alternative decryption of the ciphertext, however this is a computationally hard problem.

  7. Performance of the air2stream model that relates air and stream water temperatures depends on the calibration method

    Science.gov (United States)

    Piotrowski, Adam P.; Napiorkowski, Jaroslaw J.

    2018-06-01

    A number of physical or data-driven models have been proposed to evaluate stream water temperatures based on hydrological and meteorological observations. However, physical models require a large amount of information that is frequently unavailable, while data-based models ignore the physical processes. Recently the air2stream model has been proposed as an intermediate alternative that is based on physical heat budget processes, but it is so simplified that the model may be applied like data-driven ones. However, the price for simplicity is the need to calibrate eight parameters that, although have some physical meaning, cannot be measured or evaluated a priori. As a result, applicability and performance of the air2stream model for a particular stream relies on the efficiency of the calibration method. The original air2stream model uses an inefficient 20-year old approach called Particle Swarm Optimization with inertia weight. This study aims at finding an effective and robust calibration method for the air2stream model. Twelve different optimization algorithms are examined on six different streams from northern USA (states of Washington, Oregon and New York), Poland and Switzerland, located in both high mountains, hilly and lowland areas. It is found that the performance of the air2stream model depends significantly on the calibration method. Two algorithms lead to the best results for each considered stream. The air2stream model, calibrated with the chosen optimization methods, performs favorably against classical streamwater temperature models. The MATLAB code of the air2stream model and the chosen calibration procedure (CoBiDE) are available as Supplementary Material on the Journal of Hydrology web page.

  8. Workflow-Oriented Cyberinfrastructure for Sensor Data Analytics

    Science.gov (United States)

    Orcutt, J. A.; Rajasekar, A.; Moore, R. W.; Vernon, F.

    2015-12-01

    Sensor streams comprise an increasingly large part of Earth Science data. Analytics based on sensor data require an easy way to perform operations such as acquisition, conversion to physical units, metadata linking, sensor fusion, analysis and visualization on distributed sensor streams. Furthermore, embedding real-time sensor data into scientific workflows is of growing interest. We have implemented a scalable networked architecture that can be used to dynamically access packets of data in a stream from multiple sensors, and perform synthesis and analysis across a distributed network. Our system is based on the integrated Rule Oriented Data System (irods.org), which accesses sensor data from the Antelope Real Time Data System (brtt.com), and provides virtualized access to collections of data streams. We integrate real-time data streaming from different sources, collected for different purposes, on different time and spatial scales, and sensed by different methods. iRODS, noted for its policy-oriented data management, brings to sensor processing features and facilities such as single sign-on, third party access control lists ( ACLs), location transparency, logical resource naming, and server-side modeling capabilities while reducing the burden on sensor network operators. Rich integrated metadata support also makes it straightforward to discover data streams of interest and maintain data provenance. The workflow support in iRODS readily integrates sensor processing into any analytical pipeline. The system is developed as part of the NSF-funded Datanet Federation Consortium (datafed.org). APIs for selecting, opening, reaping and closing sensor streams are provided, along with other helper functions to associate metadata and convert sensor packets into NetCDF and JSON formats. Near real-time sensor data including seismic sensors, environmental sensors, LIDAR and video streams are available through this interface. A system for archiving sensor data and metadata in Net

  9. Quality of streams in Johnson County, Kansas, 2002--10

    Science.gov (United States)

    Rasmussen, Teresa J.; Stone, Mandy S.; Poulton, Barry C.; Graham, Jennifer L.

    2012-01-01

    Stream quality in Johnson County, northeastern Kansas, was assessed on the basis of land use, hydrology, stream-water and streambed-sediment chemistry, riparian and in-stream habitat, and periphyton and macroinvertebrate community data collected from 22 sites during 2002 through 2010. Stream conditions at the end of the study period are evaluated and compared to previous years, stream biological communities and physical and chemical conditions are characterized, streams are described relative to Kansas Department of Health and Environment impairment categories and water-quality standards, and environmental factors that most strongly correlate with biological stream quality are evaluated. The information is useful for improving water-quality management programs, documenting changing conditions with time, and evaluating compliance with water-quality standards, total maximum daily loads (TMDLs), National Pollutant Discharge Elimination System (NPDES) permit conditions, and other established guidelines and goals. Constituent concentrations in water during base flow varied across the study area and 2010 conditions were not markedly different from those measured in 2003, 2004, and 2007. Generally the highest specific conductance and concentrations of dissolved solids and major ions in water occurred at urban sites except the upstream Cedar Creek site, which is rural and has a large area of commercial and industrial land less than 1 mile upstream on both sides of the creek. The highest base-flow nutrient concentrations in water occurred downstream from wastewater treatment facilities. Water chemistry data represent base-flow conditions only, and do not show the variability in concentrations that occurs during stormwater runoff. Constituent concentrations in streambed sediment also varied across the study area and some notable changes occurred from previously collected data. High organic carbon and nutrient concentrations at the rural Big Bull Creek site in 2003 decreased

  10. A Critical Examination of the Introduction of Drug Detection Dogs for Policing of Illicit Drugs in New South Wales, Australia Using Kingdon's "Multiple Streams" Heuristic

    Science.gov (United States)

    Lancaster, Kari; Ritter, Alison; Hughes, Caitlin; Hoppe, Robert

    2017-01-01

    This paper critically analyses the introduction of drug detection dogs as a tool for policing of illicit drugs in New South Wales, Australia. Using Kingdon's "multiple streams" heuristic as a lens for analysis, we identify how the issue of drugs policing became prominent on the policy agenda, and the conditions under which the…

  11. Stream Evaluation

    Data.gov (United States)

    Kansas Data Access and Support Center — Digital representation of the map accompanying the "Kansas stream and river fishery resource evaluation" (R.E. Moss and K. Brunson, 1981.U.S. Fish and Wildlife...

  12. Sampling, Splitting and Merging in Coinductive Stream Calculus

    NARCIS (Netherlands)

    M. Niqui (Milad); J.J.M.M. Rutten (Jan); C. Bolduc; J. Desharnais; B. Ktari

    2010-01-01

    textabstractWe study various operations for partitioning, projecting and merging streams of data. These operations are motivated by their use in dataflow programming and the stream processing languages. We use the framework of \\emph{stream calculus} and \\emph{stream circuits} for defining and

  13. Sampling, splitting and merging in coinductive stream calculus

    NARCIS (Netherlands)

    M. Niqui (Milad); J.J.M.M. Rutten (Jan)

    2009-01-01

    htmlabstractWe study various operations for partitioning, projecting and merging streams of data. These operations are motivated by their use in dataflow programming and the stream processing languages. We use the framework of stream calculus and stream circuits for defining and proving properties

  14. SPEAR indicates pesticide effects in streams - Comparative use of species- and family-level biomonitoring data

    International Nuclear Information System (INIS)

    Beketov, M.A.; Foit, K.; Schaefer, R.B.; Schriever, C.A.; Sacchi, A.; Capri, E.; Biggs, J.; Wells, C.; Liess, M.

    2009-01-01

    To detect effects of pesticides on non-target freshwater organisms the Species at risk (SPEAR pesticides ) bioindicator based on biological traits was previously developed and successfully validated over different biogeographical regions of Europe using species-level data on stream invertebrates. Since many freshwater biomonitoring programmes have family-level taxonomic resolution we tested the applicability of SPEAR pesticides with family-level biomonitoring data to indicate pesticide effects in streams (i.e. insecticide toxicity of pesticides). The study showed that the explanatory power of the family-level SPEAR(fm) pesticides is not significantly lower than the species-level index. The results suggest that the family-level SPEAR(fm) pesticides is a sensitive, cost-effective, and potentially European-wide bioindicator of pesticide contamination in flowing waters. Class boundaries for SPEAR pesticides according to EU Water Framework Directive are defined to contribute to the assessment of ecological status of water bodies. - We show that SPEAR pesticides can be based on family-level biomonitoring data and is applicable for large-scale monitoring programmes to detect and quantify pesticide contamination.

  15. SPEAR indicates pesticide effects in streams - Comparative use of species- and family-level biomonitoring data

    Energy Technology Data Exchange (ETDEWEB)

    Beketov, M.A., E-mail: mikhail.beketov@ufz.d [UFZ - Helmholtz Centre for Environmental Research, Department System Ecotoxicology, Permoserstrasse 15, D-04318 Leipzig (Germany); Foit, K.; Schaefer, R.B.; Schriever, C.A. [UFZ - Helmholtz Centre for Environmental Research, Department System Ecotoxicology, Permoserstrasse 15, D-04318 Leipzig (Germany); Sacchi, A.; Capri, E. [Universita Cattolica del Sacro Cuore, Istituto di Chimica Agraria ed Ambientale, Piacenza (Italy); Biggs, J. [Pond Conservation, c/o Oxford Brookes University, Headington (United Kingdom); Wells, C. [Environment Agency of England and Wales, Science Department, Bristol (United Kingdom); Liess, M. [UFZ - Helmholtz Centre for Environmental Research, Department System Ecotoxicology, Permoserstrasse 15, D-04318 Leipzig (Germany)

    2009-06-15

    To detect effects of pesticides on non-target freshwater organisms the Species at risk (SPEAR{sub pesticides}) bioindicator based on biological traits was previously developed and successfully validated over different biogeographical regions of Europe using species-level data on stream invertebrates. Since many freshwater biomonitoring programmes have family-level taxonomic resolution we tested the applicability of SPEAR{sub pesticides} with family-level biomonitoring data to indicate pesticide effects in streams (i.e. insecticide toxicity of pesticides). The study showed that the explanatory power of the family-level SPEAR(fm){sub pesticides} is not significantly lower than the species-level index. The results suggest that the family-level SPEAR(fm){sub pesticides} is a sensitive, cost-effective, and potentially European-wide bioindicator of pesticide contamination in flowing waters. Class boundaries for SPEAR{sub pesticides} according to EU Water Framework Directive are defined to contribute to the assessment of ecological status of water bodies. - We show that SPEAR{sub pesticides} can be based on family-level biomonitoring data and is applicable for large-scale monitoring programmes to detect and quantify pesticide contamination.

  16. Comprehensive Protection of Data-Partitioned Video for Broadband Wireless IPTV Streaming

    Directory of Open Access Journals (Sweden)

    Laith Al-Jobouri

    2012-01-01

    Full Text Available This paper examines the threat to video streaming from slow and fast fading, traffic congestion, and channel packet drops. The proposed response is a combination of: rateless channel coding, which is adaptively applied; data-partitioned source coding to exploit prioritized packetization; and duplicate slice provision, which is the focus of the evaluation in this paper. The paper also considers the distribution of intra-refresh macroblocks as a means of avoiding sudden data rate increases. When error bursts occur, this paper shows that duplicate slices are certainly necessary but this provision is more effective for medium quality video than it is for high quality video. The percentage of intra-refresh macroblocks can be low and still reduce the impact of temporal error propagation.

  17. Comparison of animated jet stream visualizations

    Science.gov (United States)

    Nocke, Thomas; Hoffmann, Peter

    2016-04-01

    The visualization of 3D atmospheric phenomena in space and time is still a challenging problem. In particular, multiple solutions of animated jet stream visualizations have been produced in recent years, which were designed to visually analyze and communicate the jet and related impacts on weather circulation patterns and extreme weather events. This PICO integrates popular and new jet animation solutions and inter-compares them. The applied techniques (e.g. stream lines or line integral convolution) and parametrizations (color mapping, line lengths) are discussed with respect to visualization quality criteria and their suitability for certain visualization tasks (e.g. jet patterns and jet anomaly analysis, communicating its relevance for climate change).

  18. Quantifying Forested Riparian Buffer Ability to Ameliorate Stream Temperature in a Missouri Ozark Border Stream of the Central U.S

    Science.gov (United States)

    Bulliner, E. A.; Hubbart, J. A.

    2009-12-01

    Riparian buffers play an important role in modulating stream water quality, including temperature. There is a need to better understand riparian form and function to validate and improve contemporary management practices. Further studies are warranted to characterize energy attenuation by forested riparian canopy layers that normally buffer stream temperature, particularly in the central hardwood forest regions of the United States where relationships between canopy density and stream temperature are unknown. To quantify these complex processes, two intensively instrumented hydroclimate stations were installed along two stream reaches of a riparian stream in central Missouri, USA in the winter of 2008. Hydroclimate stations are located along stream reaches oriented in both cardinal directions, which will allow interpolation of results to other orientations. Each station consists of an array of instrumentation that senses the flux of water and energy into and out of the riparian zone. Reference data are supplied from a nearby flux tower (US DOE) located on top of a forested ridge. The study sites are located within a University of Missouri preserved wildland area on the border of the southern Missouri’s Ozark region, an ecologically distinct region in the central United States. Limestone underlies the study area, resulting in a distinct semi-Karst hydrologic system. Vegetation forms a complex, multi-layered canopy extending from the stream edge through the riparian zone and into surrounding hills. Climate is classified as humid continental, with approximate average annual temperature and precipitation of 13.2°C and 970mm, respectively. Preliminary results (summer 2009 data) indicate incoming short-wave radiation is 24.9% higher at the N-S oriented stream reach relative to the E-W oriented reach. Maximum incoming short wave radiation during the period was 64.5% lower at the N-S reach relative to E-W reach. Average air temperature for the E-W reach was 0.3°C lower

  19. Data matching for free-surface multiple attenuation by multidimensional deconvolution

    Science.gov (United States)

    van der Neut, Joost; Frijlink, Martijn; van Borselen, Roald

    2012-09-01

    A common strategy for surface-related multiple elimination of seismic data is to predict multiples by a convolutional model and subtract these adaptively from the input gathers. Problems can be posed by interfering multiples and primaries. Removing multiples by multidimensional deconvolution (MDD) (inversion) does not suffer from these problems. However, this approach requires data to be consistent, which is often not the case, especially not at interpolated near-offsets. A novel method is proposed to improve data consistency prior to inversion. This is done by backpropagating first-order multiples with a time-gated reference primary event and matching these with early primaries in the input gather. After data matching, multiple elimination by MDD can be applied with a deterministic inversion scheme.

  20. The study of Kruskal's and Prim's algorithms on the Multiple Instruction and Single Data stream computer system

    Directory of Open Access Journals (Sweden)

    A. Yu. Popov

    2015-01-01

    Full Text Available Bauman Moscow State Technical University is implementing a project to develop operating principles of computer system having radically new architecture. A developed working model of the system allowed us to evaluate an efficiency of developed hardware and software. The experimental results presented in previous studies, as well as the analysis of operating principles of new computer system permit to draw conclusions regarding its efficiency in solving discrete optimization problems related to processing of sets.The new architecture is based on a direct hardware support of operations of discrete mathematics, which is reflected in using the special facilities for processing of sets and data structures. Within the framework of the project a special device was designed, i.e. a structure processor (SP, which improved the performance, without limiting the scope of applications of such a computer system.The previous works presented the basic principles of the computational process organization in MISD (Multiple Instructions, Single Data system, showed the structure and features of the structure processor and the general principles to solve discrete optimization problems on graphs.This paper examines two search algorithms of the minimum spanning tree, namely Kruskal's and Prim's algorithms. It studies the implementations of algorithms for two SP operation modes: coprocessor mode and MISD one. The paper presents results of experimental comparison of MISD system performance in coprocessor mode with mainframes.

  1. ACTION-SPACE CLUSTERING OF TIDAL STREAMS TO INFER THE GALACTIC POTENTIAL

    Energy Technology Data Exchange (ETDEWEB)

    Sanderson, Robyn E.; Helmi, Amina [Kapteyn Astronomical Institute, P.O. Box 800, 9700 AV Groningen (Netherlands); Hogg, David W., E-mail: robyn@astro.columbia.edu [Center for Cosmology and Particle Physics, Department of Physics, New York University, 4 Washington Place, New York, NY 10003 (United States)

    2015-03-10

    We present a new method for constraining the Milky Way halo gravitational potential by simultaneously fitting multiple tidal streams. This method requires three-dimensional positions and velocities for all stars to be fit, but does not require identification of any specific stream or determination of stream membership for any star. We exploit the principle that the action distribution of stream stars is most clustered when the potential used to calculate the actions is closest to the true potential. Clustering is quantified with the Kullback-Leibler Divergence (KLD), which also provides conditional uncertainties for our parameter estimates. We show, for toy Gaia-like data in a spherical isochrone potential, that maximizing the KLD of the action distribution relative to a smoother distribution recovers the input potential. The precision depends on the observational errors and number of streams; using K III giants as tracers, we measure the enclosed mass at the average radius of the sample stars accurate to 3% and precise to 20%-40%. Recovery of the scale radius is precise to 25%, biased 50% high by the small galactocentric distance range of stars in our mock sample (1-25 kpc, or about three scale radii, with mean 6.5 kpc). 20-25 streams with at least 100 stars each are required for a stable confidence interval. With radial velocities (RVs) to 100 kpc, all parameters are determined with ∼10% accuracy and 20% precision (1.3% accuracy for the enclosed mass), underlining the need to complete the RV catalog for faint halo stars observed by Gaia.

  2. Hydrogeochemical and stream sediment reconnaissance basic data for Dodge City NTMS Quadrangle, Kansas

    International Nuclear Information System (INIS)

    1980-01-01

    Results of a reconnaissance geochemical survey of the Dodge City Quadrangle are reported. Field and laboratory data are presented for 756 groundwater and 321 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided, and pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Groundwater data indicate that the most promising areas for uranium mineralization are as follows: (1) in the north central area of the quadrangle within close proximity to the Arkansas River, mostly from waters of the Ogallala Formation; (2) in the west central area, from groundwater samples of the Dakota and the Ogallala Formations; and (3) between the North Fork of the Cimarron River and the main Cimarron River, mostly in waters from the Ogallala Formation. Associated with the high uranium values are high concentrations for magnesium, strontium, and sulfate. Of the groundwater samples taken 81% were collected from the Ogallala Formation. Stream sediment data indicate high uranium concentrations in scattered samples in the northwestern, central, and southwestern areas of the quadrangle. Most of the samples with high uranium values were collected from the Quaternary alluvium. Associated with the high uranium values are high concentrations of barium, cerium, iron, manganese, titanium, vanadium, yttrium, and zirconium

  3. Modeling transient streaming potentials in falling-head permeameter tests.

    Science.gov (United States)

    Malama, Bwalya; Revil, André

    2014-01-01

    We present transient streaming potential data collected during falling-head permeameter tests performed on samples of two sands with different physical and chemical properties. The objective of the work is to estimate hydraulic conductivity (K) and the electrokinetic coupling coefficient (Cl ) of the sand samples. A semi-empirical model based on the falling-head permeameter flow model and electrokinetic coupling is used to analyze the streaming potential data and to estimate K and Cl . The values of K estimated from head data are used to validate the streaming potential method. Estimates of K from streaming potential data closely match those obtained from the associated head data, with less than 10% deviation. The electrokinetic coupling coefficient was estimated from streaming potential vs. (1) time and (2) head data for both sands. The results indicate that, within limits of experimental error, the values of Cl estimated by the two methods are essentially the same. The results of this work demonstrate that a temporal record of the streaming potential response in falling-head permeameter tests can be used to estimate both K and Cl . They further indicate the potential for using transient streaming potential data as a proxy for hydraulic head in hydrogeology applications. © 2013, National Ground Water Association.

  4. ATLAS Live: Collaborative Information Streams

    CERN Document Server

    Goldfarb, S; The ATLAS collaboration

    2011-01-01

    I report on a pilot project launched in 2010 focusing on facilitating communication and information exchange within the ATLAS Collaboration, through the combination of digital signage software and webcasting. The project, called ATLAS Live, implements video streams of information, ranging from detailed detector and data status to educational and outreach material. The content, including text, images, video and audio, is collected, visualised and scheduled using digital signage software. The system is robust and flexible, utilizing scripts to input data from remote sources, such as the CERN Document Server, Indico, or any available URL, and to integrate these sources into professional-quality streams, including text scrolling, transition effects, inter and intra-screen divisibility. Information is published via the encoding and webcasting of standard video streams, viewable on all common platforms, using a web browser or other common video tool. Authorisation is enforced at the level of the streaming and at th...

  5. Engineering hyporheic zones to target nitrification versus denitrification: performance data from constructed stream flumes

    Science.gov (United States)

    Herzog, S.; Portmann, A. C.; Halpin, B. N.; Higgins, C.; McCray, J. E.

    2017-12-01

    Nonpoint source nitrogen pollution from agricultural and urban runoff is one of the leading causes of impairment to US rivers and streams. The hyporheic zone (HZ) offers a natural biogeochemical hotspot for the attenuation of nitrogen within streams, thereby complementing efforts to prevent aquatic nitrogen pollution in the first place. However, HZ in urban and agricultural streams are often degraded by scouring and colmation, which limit their potential to improve stream water quality at the reach scale. A recent effort to mitigate nitrogen pollution in the Chesapeake Bay region provides denitrification credits for hyporheic restoration projects. Unfortunately, many of the featured hyporheic zone best management practices (BMP) (e.g., weirs, cross-vanes) tend to create only localized, aerobic hyporheic flows that are not optimal for the anaerobic denitrification reaction. In short, practitioners lack an adaptable BMP that can both 1) increase hyporheic exchange, and 2) tailor HZ residence times to match reactions of interest. Here we present new performance data for an HZ engineering technique called Biohydrochemical Enhancements for Streamwater Treatment (BEST). BEST are subsurface modules that utilize low-permeability sediments to drive efficient hyporheic exchange and control residence times, along with reactive geomedia to increase reaction rates within HZ sediments. This research utilized two artificial stream flumes: One flume served as an all-sand control condition, the other featured BEST modules at 1m spacing with a mixture of 70/30 sand/woodchips (v/v). Two different BEST media were tested: a coarse sand module with K 0.5 cm/s, and a fine sand module with K 0.15 cm/s. The flume with coarse sand BEST modules created aerobic HZ conditions and demonstrated rapid nitrification of ammonia at rates significantly higher than the control. However, denitrification was much slower and not significantly different between the two streams. In contrast, the fine sand

  6. Multiple Kernel Learning for Heterogeneous Anomaly Detection: Algorithm and Aviation Safety Case Study

    Science.gov (United States)

    Das, Santanu; Srivastava, Ashok N.; Matthews, Bryan L.; Oza, Nikunj C.

    2010-01-01

    The world-wide aviation system is one of the most complex dynamical systems ever developed and is generating data at an extremely rapid rate. Most modern commercial aircraft record several hundred flight parameters including information from the guidance, navigation, and control systems, the avionics and propulsion systems, and the pilot inputs into the aircraft. These parameters may be continuous measurements or binary or categorical measurements recorded in one second intervals for the duration of the flight. Currently, most approaches to aviation safety are reactive, meaning that they are designed to react to an aviation safety incident or accident. In this paper, we discuss a novel approach based on the theory of multiple kernel learning to detect potential safety anomalies in very large data bases of discrete and continuous data from world-wide operations of commercial fleets. We pose a general anomaly detection problem which includes both discrete and continuous data streams, where we assume that the discrete streams have a causal influence on the continuous streams. We also assume that atypical sequence of events in the discrete streams can lead to off-nominal system performance. We discuss the application domain, novel algorithms, and also discuss results on real-world data sets. Our algorithm uncovers operationally significant events in high dimensional data streams in the aviation industry which are not detectable using state of the art methods

  7. Solar wind stream interfaces

    International Nuclear Information System (INIS)

    Gosling, J.T.; Asbridge, J.R.; Bame, S.J.; Feldman, W.C.

    1978-01-01

    Measurements aboard Imp 6, 7, and 8 reveal that approximately one third of all high-speed solar wind streams observed at 1 AU contain a sharp boundary (of thickness less than approx.4 x 10 4 km) near their leading edge, called a stream interface, which separates plasma of distinctly different properties and origins. Identified as discontinuities across which the density drops abruptly, the proton temperature increases abruptly, and the speed rises, stream interfaces are remarkably similar in character from one stream to the next. A superposed epoch analysis of plasma data has been performed for 23 discontinuous stream interfaces observed during the interval March 1971 through August 1974. Among the results of this analysis are the following: (1) a stream interface separates what was originally thick (i.e., dense) slow gas from what was originally thin (i.e., rare) fast gas; (2) the interface is the site of a discontinuous shear in the solar wind flow in a frame of reference corotating with the sun; (3) stream interfaces occur at speeds less than 450 km s - 1 and close to or at the maximum of the pressure ridge at the leading edges of high-speed streams; (4) a discontinuous rise by approx.40% in electron temperature occurs at the interface; and (5) discontinuous changes (usually rises) in alpha particle abundance and flow speed relative to the protons occur at the interface. Stream interfaces do not generally recur on successive solar rotations, even though the streams in which they are embedded often do. At distances beyond several astronomical units, stream interfaces should be bounded by forward-reverse shock pairs; three of four reverse shocks observed at 1 AU during 1971--1974 were preceded within approx.1 day by stream interfaces. Our observations suggest that many streams close to the sun are bounded on all sides by large radial velocity shears separating rapidly expanding plasma from more slowly expanding plasma

  8. PGG: An Online Pattern Based Approach for Stream Variation Management

    Institute of Scientific and Technical Information of China (English)

    Lu-An Tang; Bin Cui; Hong-Yan Li; Gao-Shan Miao; Dong-Qing Yang; Xin-Biao Zhou

    2008-01-01

    Many database applications require efficient processing of data streams with value variations and fiuctuant sampling frequency. The variations typically imply fundamental features of the stream and important domain knowledge of underlying objects. In some data streams, successive events seem to recur in a certain time interval, but the data indeed evolves with tiny differences as time elapses. This feature, so called pseudo periodicity, poses a new challenge to stream variation management. This study focuses on the online management for variations over such streams. The idea can be applied to many scenarios such as patient vital signal monitoring in medical applications. This paper proposes a new method named Pattern Growth Graph (PGG) to detect and manage variations over evolving streams with following features: 1) adopts the wave-pattern to capture the major information of data evolution and represent them compactly;2) detects the variations in a single pass over the stream with the help of wave-pattern matching algorithm; 3) only stores different segments of the pattern for incoming stream, and hence substantially compresses the data without losing important information; 4) distinguishes meaningful data changes from noise and reconstructs the stream with acceptable accuracy.Extensive experiments on real datasets containing millions of data items, as well as a prototype system, are carried out to demonstrate the feasibility and effectiveness of the proposed scheme.

  9. Multiple Kernel Learning with Data Augmentation

    Science.gov (United States)

    2016-11-22

    JMLR: Workshop and Conference Proceedings 63:49–64, 2016 ACML 2016 Multiple Kernel Learning with Data Augmentation Khanh Nguyen nkhanh@deakin.edu.au...University, Australia Editors: Robert J. Durrant and Kee-Eung Kim Abstract The motivations of multiple kernel learning (MKL) approach are to increase... kernel expres- siveness capacity and to avoid the expensive grid search over a wide spectrum of kernels . A large amount of work has been proposed to

  10. Towards automatic parameter tuning of stream processing systems

    KAUST Repository

    Bilal, Muhammad; Canini, Marco

    2017-01-01

    for automating parameter tuning for stream-processing systems. Our framework supports standard black-box optimization algorithms as well as a novel gray-box optimization algorithm. We demonstrate the multiple benefits of automated parameter tuning in optimizing

  11. Round-Robin Streaming with Generations

    DEFF Research Database (Denmark)

    Li, Yao; Vingelmann, Peter; Pedersen, Morten Videbæk

    2012-01-01

    We consider three types of application layer coding for streaming over lossy links: random linear coding, systematic random linear coding, and structured coding. The file being streamed is divided into sub-blocks (generations). Code symbols are formed by combining data belonging to the same...

  12. HIDE AND SEEK BETWEEN ANDROMEDA'S HALO, DISK, AND GIANT STREAM

    Energy Technology Data Exchange (ETDEWEB)

    Clementini, Gisella; Contreras Ramos, Rodrigo; Federici, Luciana; Macario, Giulia; Tosi, Monica; Bellazzini, Michele; Fusi Pecci, Flavio; Diolaiti, Emiliano; Cacciari, Carla [INAF, Osservatorio Astronomico di Bologna, Bologna (Italy); Beccari, Giacomo [European Southern Observatory, 85748 Garching bei Munchen (Germany); Testa, Vincenzo; Giallongo, Emanuele; Di Paola, Andrea; Gallozzi, Stefano [INAF, Osservatorio Astronomico di Roma, Monteporzio (Italy); Cignoni, Michele; Marano, Bruno [Dipartimento di Astronomia, Universita di Bologna, Bologna (Italy); Marconi, Marcella; Ripepi, Vincenzo [INAF, Osservatorio Astronomico di Capodimonte, Napoli (Italy); Ragazzoni, Roberto [INAF, Osservatorio Astronomico di Padova, Padova (Italy); Smareglia, Riccardo, E-mail: gisella.clementini@oabo.inaf.it [INAF, Osservatorio Astronomico di Trieste, Trieste (Italy)

    2011-12-10

    Photometry in B, V (down to V {approx} 26 mag) is presented for two 23' Multiplication-Sign 23' fields of the Andromeda galaxy (M31) that were observed with the blue channel camera of the Large Binocular Telescope during the Science Demonstration Time. Each field covers an area of about 5.1 Multiplication-Sign 5.1 kpc{sup 2} at the distance of M31 ({mu}{sub M31} {approx} 24.4 mag), sampling, respectively, a northeast region close to the M31 giant stream (field S2) and an eastern portion of the halo in the direction of the galaxy minor axis (field H1). The stream field spans a region that includes Andromeda's disk and giant stream, and this is reflected in the complexity of the color-magnitude diagram of the field. One corner of the halo field also includes a portion of the giant stream. Even though these demonstration time data were obtained under non-optimal observing conditions, the B photometry, which was acquired in time-series mode, allowed us to identify 274 variable stars (among which 96 are bona fide and 31 are candidate RR Lyrae stars, 71 are Cepheids, and 16 are binary systems) by applying the image subtraction technique to the selected portions of the observed fields. Differential flux light curves were obtained for the vast majority of these variables. Our sample mainly includes pulsating stars that populate the instability strip from the Classical Cepheids down to the RR Lyrae stars, thus tracing the different stellar generations in these regions of M31 down to the horizontal branch of the oldest (t {approx} 10 Gyr) component.

  13. Salamander occupancy in headwater stream networks

    Science.gov (United States)

    Grant, E.H.C.; Green, L.E.; Lowe, W.H.

    2009-01-01

    1. Stream ecosystems exhibit a highly consistent dendritic geometry in which linear habitat units intersect to create a hierarchical network of connected branches. 2. Ecological and life history traits of species living in streams, such as the potential for overland movement, may interact with this architecture to shape patterns of occupancy and response to disturbance. Specifically, large-scale habitat alteration that fragments stream networks and reduces connectivity may reduce the probability a stream is occupied by sensitive species, such as stream salamanders. 3. We collected habitat occupancy data on four species of stream salamanders in first-order (i.e. headwater) streams in undeveloped and urbanised regions of the eastern U.S.A. We then used an information-theoretic approach to test alternative models of salamander occupancy based on a priori predictions of the effects of network configuration, region and salamander life history. 4. Across all four species, we found that streams connected to other first-order streams had higher occupancy than those flowing directly into larger streams and rivers. For three of the four species, occupancy was lower in the urbanised region than in the undeveloped region. 5. These results demonstrate that the spatial configuration of stream networks within protected areas affects the occurrences of stream salamander species. We strongly encourage preservation of network connections between first-order streams in conservation planning and management decisions that may affect stream species.

  14. Meteorological, stream-discharge, and water-quality data for water year 1992 from two basins in Central Nevada

    International Nuclear Information System (INIS)

    McKinley, P.W.; Oliver, T.A.

    1995-01-01

    The US Geological Survey, in cooperation with the US Department of Energy, is studying Yucca Mountain, Nevada, as a potential repository for high level nuclear waste. As part of the Yucca Mountain Site Project, the analog recharge study is providing data for the evaluation of recharge to the Yucca Mountain ground-water system given a cooler and wetter climate than currently exists. The current and climatic conditions are favorable to the isolation of radioactive waste. Because waste isolation from the accessible environment for 10,000 years is necessary, climatic change and the potential for increased ground-water recharge need to be considered as part of the characterization of the potential repository. Therefore, two small basins, measuring less than 2 square miles, were studied to determine the volume of precipitation available for recharge to ground water. The semiarid 3-Springs Basin is located to the east of Kawich Peak in the Kawich Range east of Tonopah, Nevada. Stewart Basin is a subalpine drainage basin north of Arc Dome in the Toiyabe Range north of Tonopah, Nevada. The purpose of this publication is to make available the meteorological, stream-discharge, and water-quality data collected during the study. Meteorological data collected include air temperature, soil temperature, solar radiation, and relative humidity. Stream-discharge data were collected from the surface-water outlet of each basin. Water-quality data are chemical analyses of water samples collected from surface- and ground-water sources. Each basin has a meteorological station located in the lower and upper reaches of the basin. Hydrologic records include stream-discharge and water-quality data from the lower meteorological site and water-quality data from springs within the basins

  15. Stream Habitat Reach Summary - North Coast [ds63

    Data.gov (United States)

    California Natural Resource Agency — The shapefile is based on habitat unit level data summarized at the stream reach level. The database represents salmonid stream habitat surveys from 645 streams of...

  16. Computationally efficient implementation of sarse-tap FIR adaptive filters with tap-position control on intel IA-32 processors

    OpenAIRE

    Hirano, Akihiro; Nakayama, Kenji

    2008-01-01

    This paper presents an computationally ef cient implementation of sparse-tap FIR adaptive lters with tapposition control on Intel IA-32 processors with single-instruction multiple-data (SIMD) capability. In order to overcome randomorder memory access which prevents a ectorization, a blockbased processing and a re-ordering buffer are introduced. A dynamic register allocation and the use of memory-to-register operations help the maximization of the loop-unrolling level. Up to 66percent speedup ...

  17. Hydrogeochemical and stream sediment reconnaissance basic data for Watertown NTMS Quadrangle, South Dakota; Minnesota

    International Nuclear Information System (INIS)

    1981-01-01

    Results of a reconnaissance geochemical survey of the Watertown Quadrangle are reported. Field and laboratory data are presented for 711 groundwater and 603 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided, and pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Groundwater data indicate that high uranium concentrations are derived predominantly from glacial aquifers of variable water composition located on the Coteau des Prairies. Elements associated with high uranium values in these waters include barium, calcium, copper, iron, magnesium, selenium, sulfate, and total alkalinity. Low uranium values were observed in waters originating from the Cretaceous Dakota sandstone whose water chemistry is characterized by high concentrations of boron, sodium, and chloride. Stream sediment data indicate that high uranium concentrations are scattered across the glacial deposits of the Coteau des Prairies. A major clustering of high uranium values occurs in the eastern portion of the glaciated quadrangle and is associated with high concentrations of selenium, lithium, iron, arsenic, chromium, and vanadium. The sediment data suggest that the drift covering the Watertown Quadrangle is compositionally homogeneous, although subtle geochemical differences were observed as a result of localized contrasts in drift source-rock mineralogy and modification of elemental distributions by contemporaneous and postglacial hydrologic processes

  18. The Effectiveness of Streaming Video on Medical Student Learning: A Case Study

    OpenAIRE

    Bridge, Patrick D.; Jackson, Matt; Robinson, Leah

    2009-01-01

    Information technology helps meet today’s medical students’ needs by providing multiple curriculum delivery methods. Video streaming is an e-learning technology that uses the Internet to deliver curriculum while giving the student control of the content’s delivery. There have been few studies conducted on the effectiveness of streaming video in medical schools. A 5-year retrospective study was conducted using three groups of students (n_1736) to determine if the availability of streaming vide...

  19. Streaming Visual Analytics Workshop Report

    Energy Technology Data Exchange (ETDEWEB)

    Cook, Kristin A. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Burtner, Edwin R. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Kritzstein, Brian P. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Brisbois, Brooke R. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Mitson, Anna E. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2016-03-31

    How can we best enable users to understand complex emerging events and make appropriate assessments from streaming data? This was the central question addressed at a three-day workshop on streaming visual analytics. This workshop was organized by Pacific Northwest National Laboratory for a government sponsor. It brought together forty researchers and subject matter experts from government, industry, and academia. This report summarizes the outcomes from that workshop. It describes elements of the vision for a streaming visual analytic environment and set of important research directions needed to achieve this vision. Streaming data analysis is in many ways the analysis and understanding of change. However, current visual analytics systems usually focus on static data collections, meaning that dynamically changing conditions are not appropriately addressed. The envisioned mixed-initiative streaming visual analytics environment creates a collaboration between the analyst and the system to support the analysis process. It raises the level of discourse from low-level data records to higher-level concepts. The system supports the analyst’s rapid orientation and reorientation as situations change. It provides an environment to support the analyst’s critical thinking. It infers tasks and interests based on the analyst’s interactions. The system works as both an assistant and a devil’s advocate, finding relevant data and alerts as well as considering alternative hypotheses. Finally, the system supports sharing of findings with others. Making such an environment a reality requires research in several areas. The workshop discussions focused on four broad areas: support for critical thinking, visual representation of change, mixed-initiative analysis, and the use of narratives for analysis and communication.

  20. Assessing the chemical contamination dynamics in a mixed land use stream system.

    Science.gov (United States)

    Sonne, Anne Th; McKnight, Ursula S; Rønde, Vinni; Bjerg, Poul L

    2017-11-15

    Traditionally, the monitoring of streams for chemical and ecological status has been limited to surface water concentrations, where the dominant focus has been on general water quality and the risk for eutrophication. Mixed land use stream systems, comprising urban areas and agricultural production, are challenging to assess with multiple chemical stressors impacting stream corridors. New approaches are urgently needed for identifying relevant sources, pathways and potential impacts for implementation of suitable source management and remedial measures. We developed a method for risk assessing chemical stressors in these systems and applied the approach to a 16-km groundwater-fed stream corridor (Grindsted, Denmark). Three methods were combined: (i) in-stream contaminant mass discharge for source quantification, (ii) Toxic Units and (iii) environmental standards. An evaluation of the chemical quality of all three stream compartments - stream water, hyporheic zone, streambed sediment - made it possible to link chemical stressors to their respective sources and obtain new knowledge about source composition and origin. Moreover, toxic unit estimation and comparison to environmental standards revealed the stream water quality was substantially impaired by both geogenic and diffuse anthropogenic sources of metals along the entire corridor, while the streambed was less impacted. Quantification of the contaminant mass discharge originating from a former pharmaceutical factory revealed that several 100 kgs of chlorinated ethenes and pharmaceutical compounds discharge into the stream every year. The strongly reduced redox conditions in the plume result in high concentrations of dissolved iron and additionally release arsenic, generating the complex contaminant mixture found in the narrow discharge zone. The fingerprint of the plume was observed in the stream several km downgradient, while nutrients, inorganics and pesticides played a minor role for the stream health. The

  1. Solute-specific patterns and drivers of urban stream chemistry revealed by long-term monitoring in Baltimore, Maryland

    Science.gov (United States)

    Reisinger, A. J.; Woytowitz, E.; Majcher, E.; Rosi, E. J.; Groffman, P.

    2017-12-01

    Urban streams receive a myriad of chemical inputs from the surrounding landscape due to altered lithology (asphalt, concrete), leaky sewage infrastructure, and other human activities (road salt, fertilizer, industrial wastes, wastewater effluent), potentially leading to multiple chemical stressors occurring simultaneously. To evaluate potential drivers of water chemistry change, we used approximately 20 years of weekly water chemistry monitoring data from streams in the Baltimore Ecosystem Study (BES) to quantify trends of annual loads and flow-weighted concentrations for multiple solutes of interest, including nitrate (NO3-), phosphate (PO43-), total nitrogen (TN), total phosphorus (TP), chloride (Cl-), and sulfate (SO42-) and subsequently examined various gray and green infrastructure characteristics at the watershed scale. For example, we quantified annual volume and duration of reported sanitary sewer overflows (SSO) and cumulative storage volume and area of various best management practices (BMPs). Site- and solute-specific trends differed, but across our monitoring network we found evidence for decreasing annual export for multiple solutes. Additionally, we found that changes in gray- and green-infrastructure characteristics were related to changes in water quality at our most downstream (most urban) monitoring site. For example, annual NO3- loads increased with longer cumulative SSO duration, whereas annual PO43- and TP loads decreased with a cumulative BMP area in the watershed. Further, we used same long-term water chemistry data and multivariate analyses to investigate whether urban streams have unique water chemistry fingerprints representing the multiple chemical stressors at a given site, which could provide insight into sources and impacts of water-quality impairment. These analyses and results illustrate the major role gray and green infrastructure play in influencing water quality in urban environments, and illustrate that focusing on a variety of

  2. Hydrogeochemical and stream sediment reconnaissance basic data report for Williams NTMS quadrangle, Arizona

    Energy Technology Data Exchange (ETDEWEB)

    Wagoner, J.L.

    1979-02-01

    Wet and dry sediments were collected throughout the 18,500-km/sup 2/arid-to-semiarid region and water samples at available streams, springs, and wells. Samples were collected between August 1977 and January 1978. Results of neutron activation analyses of uranium and trace elements and other field and laboratory analyses are presented in tabular hardcopy and microfiche format. The report includes six full-size overlays for use with the Williams NTMS 1:250,000 quadrangle. Sediment samples are divided into five general groups according to the source rock from which the sediment was derived. Background uranium concentrations for the quadrangle are relatively low, ranging from 1.91 to 2.40 ppM, with the highest associated with the Precambrian igneous and metamorphic complexes of the Basin and Range province. Uranium correlates best with the rare-earth elements and iron, scandium, titanium, and manganese. Known uranium occurrences are not readily identified by the stream sediment data.

  3. Hydrogeochemical and stream sediment reconnaissance basic data report for Williams NTMS quadrangle, Arizona

    International Nuclear Information System (INIS)

    Wagoner, J.L.

    1979-02-01

    Wet and dry sediments were collected throughout the 18,500-km 2 arid-to-semiarid region and water samples at available streams, springs, and wells. Samples were collected between August 1977 and January 1978. Results of neutron activation analyses of uranium and trace elements and other field and laboratory analyses are presented in tabular hardcopy and microfiche format. The report includes six full-size overlays for use with the Williams NTMS 1:250,000 quadrangle. Sediment samples are divided into five general groups according to the source rock from which the sediment was derived. Background uranium concentrations for the quadrangle are relatively low, ranging from 1.91 to 2.40 ppM, with the highest associated with the Precambrian igneous and metamorphic complexes of the Basin and Range province. Uranium correlates best with the rare-earth elements and iron, scandium, titanium, and manganese. Known uranium occurrences are not readily identified by the stream sediment data

  4. ATLAS Live: Collaborative Information Streams

    Energy Technology Data Exchange (ETDEWEB)

    Goldfarb, Steven [Department of Physics, University of Michigan, Ann Arbor, MI 48109 (United States); Collaboration: ATLAS Collaboration

    2011-12-23

    I report on a pilot project launched in 2010 focusing on facilitating communication and information exchange within the ATLAS Collaboration, through the combination of digital signage software and webcasting. The project, called ATLAS Live, implements video streams of information, ranging from detailed detector and data status to educational and outreach material. The content, including text, images, video and audio, is collected, visualised and scheduled using digital signage software. The system is robust and flexible, utilizing scripts to input data from remote sources, such as the CERN Document Server, Indico, or any available URL, and to integrate these sources into professional-quality streams, including text scrolling, transition effects, inter and intra-screen divisibility. Information is published via the encoding and webcasting of standard video streams, viewable on all common platforms, using a web browser or other common video tool. Authorisation is enforced at the level of the streaming and at the web portals, using the CERN SSO system.

  5. ATLAS Live: Collaborative Information Streams

    International Nuclear Information System (INIS)

    Goldfarb, Steven

    2011-01-01

    I report on a pilot project launched in 2010 focusing on facilitating communication and information exchange within the ATLAS Collaboration, through the combination of digital signage software and webcasting. The project, called ATLAS Live, implements video streams of information, ranging from detailed detector and data status to educational and outreach material. The content, including text, images, video and audio, is collected, visualised and scheduled using digital signage software. The system is robust and flexible, utilizing scripts to input data from remote sources, such as the CERN Document Server, Indico, or any available URL, and to integrate these sources into professional-quality streams, including text scrolling, transition effects, inter and intra-screen divisibility. Information is published via the encoding and webcasting of standard video streams, viewable on all common platforms, using a web browser or other common video tool. Authorisation is enforced at the level of the streaming and at the web portals, using the CERN SSO system.

  6. Hydrogeochemical and stream sediment reconnaissance basic data for Ashland NTMS Quadrangle, Wisconsin; Michigan; Minnesota

    International Nuclear Information System (INIS)

    1979-01-01

    Results of a reconnaissance geochemical survey of the Ashland Quadrangle, Wisconsin; Michigan; Minnesota are reported. Field and laboratory data are presented for 312 groundwater and 383 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. A generalized geologic map of the survey area is provided, and pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Groundwater data indicate that the most promising area for potential uranium mineralization occurs along the Douglas Thrust Fault in northern Douglas County, Wisconsin. The Douglas Fault brings Fond du Lac Formation sediments in contact with Chengwatana volcanics where carbonate-rich water derived from the mafic volcanics enter the arkosic Fond du Lac Formation. Another area of interest surrounds the Bad River Indian Reservation in northern Ashland and Iron Counties. The waters here are produced from red lithic sandstone and are also associated with the Douglas Fault. Water chemistry of these waters appears similar to the waters from the Douglas County area. The stream sediment data are inconclusive because of the extensive cover of glacial deposits. A moderately favorable area is present in a strip along Lake Superior in Douglas County, where sediments are derived from arkoses of the Fond du Lac Formation

  7. Interactive volume exploration of petascale microscopy data streams using a visualization-driven virtual memory approach

    KAUST Repository

    Hadwiger, Markus; Beyer, Johanna; Jeong, Wonki; Pfister, Hanspeter

    2012-01-01

    This paper presents the first volume visualization system that scales to petascale volumes imaged as a continuous stream of high-resolution electron microscopy images. Our architecture scales to dense, anisotropic petascale volumes because it: (1) decouples construction of the 3D multi-resolution representation required for visualization from data acquisition, and (2) decouples sample access time during ray-casting from the size of the multi-resolution hierarchy. Our system is designed around a scalable multi-resolution virtual memory architecture that handles missing data naturally, does not pre-compute any 3D multi-resolution representation such as an octree, and can accept a constant stream of 2D image tiles from the microscopes. A novelty of our system design is that it is visualization-driven: we restrict most computations to the visible volume data. Leveraging the virtual memory architecture, missing data are detected during volume ray-casting as cache misses, which are propagated backwards for on-demand out-of-core processing. 3D blocks of volume data are only constructed from 2D microscope image tiles when they have actually been accessed during ray-casting. We extensively evaluate our system design choices with respect to scalability and performance, compare to previous best-of-breed systems, and illustrate the effectiveness of our system for real microscopy data from neuroscience. © 1995-2012 IEEE.

  8. Interactive volume exploration of petascale microscopy data streams using a visualization-driven virtual memory approach

    KAUST Repository

    Hadwiger, Markus

    2012-12-01

    This paper presents the first volume visualization system that scales to petascale volumes imaged as a continuous stream of high-resolution electron microscopy images. Our architecture scales to dense, anisotropic petascale volumes because it: (1) decouples construction of the 3D multi-resolution representation required for visualization from data acquisition, and (2) decouples sample access time during ray-casting from the size of the multi-resolution hierarchy. Our system is designed around a scalable multi-resolution virtual memory architecture that handles missing data naturally, does not pre-compute any 3D multi-resolution representation such as an octree, and can accept a constant stream of 2D image tiles from the microscopes. A novelty of our system design is that it is visualization-driven: we restrict most computations to the visible volume data. Leveraging the virtual memory architecture, missing data are detected during volume ray-casting as cache misses, which are propagated backwards for on-demand out-of-core processing. 3D blocks of volume data are only constructed from 2D microscope image tiles when they have actually been accessed during ray-casting. We extensively evaluate our system design choices with respect to scalability and performance, compare to previous best-of-breed systems, and illustrate the effectiveness of our system for real microscopy data from neuroscience. © 1995-2012 IEEE.

  9. Pocatello 10 x 20 NTMS area Idaho. Data report: National Uranium Resource Evaluation program, hydrogeochemical and stream sediment reconnaissance

    International Nuclear Information System (INIS)

    Cook, J.R.

    1980-07-01

    This data report presents results of groundwater and stream/surface sediment reconnaissance in the National Topographic Map Series (NTMS) Pocatello 1 0 x 2 0 quadrangle. Surface samples (sediment) were collected from 1701 sites. The target sampling density was one site per 16 square kilometers (six square miles). Ground water samples were collected at 381 sites. Neutron activation analysis (NAA) results are given for uranium and 16 other elements in sediments, and for uranium and 9 other elements in ground water. Mass spectrometry results are given for helium in ground water. Field measurements and observations are reported for each site. Analytical data and field measurements are presented in tables and maps. Statistical summaries of data and a brief description of results are given. A generalized geologic map and a summary of the geology of the area are included. Data from sediment sites include: (1) stream water chemistry measurements where applicable (pH, conductivity, and alkalinity); and (2) elemental analyses for sediment samples (U, Th, Hf, Al, Ce, Dy, Eu, Fe, La, Lu, Mn, Sc, Sm, Na, Ti, V, and Yb). Sample site descriptors (stream characteristics, vegetation, etc.) are also tabulated. Areal distribution maps, histograms, and cumulative frequency plots for most elements; U/Th, U/Hf, and U/La ratios; and scintillometer readings for sediment sample sites are included on the microfiche. Data from groundwater sites include: (1) water chemistry measurements (pH, conductivity, and alkalinity); (2) physical measurements where applicable (water temperature, well description, and scintillometer reading); and (3) elemental analyses (U, Al, Br, Cl, Dy, F, He, Mg, Mn, Na, and V). Data from stream water sites include: (1) water chemistry measurements (pH, conductivity, and alkalinity); and (2) elemental analyses

  10. Parametric analysis of neutron streaming through major penetrations in the 0.914 m TFTR test cell floor

    International Nuclear Information System (INIS)

    Ku, L.P.; Liew, S.L.; Kolibal, J.G.

    1985-09-01

    Neutron streaming through penetrations in the 0.914 m TFTR test cell floor has two distinct features: (1) the oblique angle of incidence; and (2) the high order of anisotropy in the angular distribution for incident neutrons with energies > 10 keV. The effects of these features on the neutron streaming into the TFTR basement were studied parametrically for isolated penetrations. Variations with respect to the source energies, angular distributions, and sizes of the penetrations were made. The results form a data base from which the spatial distribution of the neutron flux in the basement due to multiple penetrations may be evaluated

  11. THE ORBIT OF THE ORPHAN STREAM

    International Nuclear Information System (INIS)

    Newberg, Heidi Jo; Willett, Benjamin A.; Yanny, Brian; Xu Yan

    2010-01-01

    We use recent Sloan Extension for Galactic Understanding and Exploration (SEGUE) spectroscopy and the Sloan Digital Sky Survey (SDSS) and SEGUE imaging data to measure the sky position, distance, and radial velocities of stars in the tidal debris stream that is commonly referred to as the 'Orphan Stream'. We fit orbital parameters to the data and find a prograde orbit with an apogalacticon, perigalacticon, and eccentricity of 90 kpc, 16.4 kpc, and e = 0.7, respectively. Neither the dwarf galaxy UMa II nor the Complex A gas cloud has velocities consistent with a kinematic association with the Orphan Stream. It is possible that Segue-1 is associated with the Orphan Stream, but no other known Galactic clusters or dwarf galaxies in the Milky Way lie along its orbit. The detected portion of the stream ranges from 19 to 47 kpc from the Sun and is an indicator of the mass interior to these distances. There is a marked increase in the density of Orphan Stream stars near (l, b) = (253 0 , 49 0 ), which could indicate the presence of the progenitor at the edge of the SDSS data. If this is the progenitor, then the detected portion of the Orphan Stream is a leading tidal tail. We find blue horizontal branch (BHB) stars and F turnoff stars associated with the Orphan Stream. The turnoff color is (g - r) 0 = 0.22. The BHB stars have a low metallicity of [Fe/H] WBG = -2.1. The orbit is best fit to a halo potential with a halo plus disk mass of about 2.6 x 10 11 M sun , integrated to 60 kpc from the Galactic center. Our fits are done to orbits rather than full N-body simulations; we show that if N-body simulations are used, the inferred mass of the galaxy would be slightly smaller. Our best fit is found with a logarithmic halo speed of v halo = 73 ± 24 km s -1 , a disk+bulge mass of M(R 11 M sun , and a halo mass of M(R 11 M sun . However, we can find similar fits to the data that use a Navarro-Frenk-White halo profile or that have smaller disk masses and correspondingly larger

  12. Relating stream function and land cover in the Middle Pee Dee River Basin, SC

    Directory of Open Access Journals (Sweden)

    A.D. Jayakaran

    2016-03-01

    Full Text Available Study region: The study region comprised sixteen stream sites and associated contributing watersheds located in the Middle Pee Dee River Basin (MPDRB of South Carolina, USA. Study focus: The study was conducted between 2008 and 2010 to quantify how indices of streamflow varied with land cover characteristics analyzed at multiple spatial scales and fluvial geomorphic characteristics of sampled streams in the MPDRB. Study objectives were to relate three indices of streamflow that reflect recent temporal flow variability in a stream, with synoptic stream geomorphological measurements, and land cover type at specific spatial domains. New hydrological insights for the region: Modifications to the landscape, hydrologic regime, and alteration to channel morphology, are major threats to the functioning of riparian ecosystem functions but can rarely be linked to a single common stressor. Results from the study showed that in the MPDRB, wetland cover in the riparian corridor was an important factor, correlating significantly with stream flashiness, channel enlargement, and bed substrate character. It was also shown that a combination of stream geomorphological characteristics when combined with landscape variables at specific spatial scales were reasonable predictors of all three indices of streamflow. The study also highlights an innovative statistical methodology to relate land cover data to commonly measured metrics of streamflow and fluvial geomorphology. Keywords: Flashiness, Stream habitat, Flow indices, Land cover analysis, Wetlands, Coastal plain, Bed material, Partial least squares regression, Pee Dee River, South Carolina

  13. THE PAL 5 STAR STREAM GAPS

    International Nuclear Information System (INIS)

    Carlberg, R. G.; Hetherington, Nathan; Grillmair, C. J.

    2012-01-01

    Pal 5 is a low-mass, low-velocity-dispersion, globular cluster with spectacular tidal tails. We use the Sloan Digital Sky Survey Data Release 8 data to extend the density measurements of the trailing star stream to 23 deg distance from the cluster, at which point the stream runs off the edge of the available sky coverage. The size and the number of gaps in the stream are measured using a filter which approximates the structure of the gaps found in stream simulations. We find 5 gaps that are at least 99% confidence detections with about a dozen gaps at 90% confidence. The statistical significance of a gap is estimated using bootstrap resampling of the control regions on either side of the stream. The density minimum closest to the cluster is likely the result of the epicyclic orbits of the tidal outflow and has been discounted. To create the number of 99% confidence gaps per unit length at the mean age of the stream requires a halo population of nearly a thousand dark matter sub-halos with peak circular velocities above 1 km s –1 within 30 kpc of the galactic center. These numbers are a factor of about three below cold stream simulation at this sub-halo mass or velocity but, given the uncertainties in both measurement and more realistic warm stream modeling, are in substantial agreement with the LCDM prediction.

  14. Computer analysis to the geochemical of soil and stream sediments data in an area of Southern Uruguay

    International Nuclear Information System (INIS)

    Spangenberg, J.

    2012-01-01

    This work is about geochemical interpretation of multi-element data from a soil and stream sediment survey carried out in Southern of Uruguay .This zone has several occurrences of metal sulphide mineralization

  15. Carbon and nitrogen stoichiometry across stream ecosystems

    Science.gov (United States)

    Wymore, A.; Kaushal, S.; McDowell, W. H.; Kortelainen, P.; Bernhardt, E. S.; Johnes, P.; Dodds, W. K.; Johnson, S.; Brookshire, J.; Spencer, R.; Rodriguez-Cardona, B.; Helton, A. M.; Barnes, R.; Argerich, A.; Haq, S.; Sullivan, P. L.; López-Lloreda, C.; Coble, A. A.; Daley, M.

    2017-12-01

    Anthropogenic activities are altering carbon and nitrogen concentrations in surface waters globally. The stoichiometry of carbon and nitrogen regulates important watershed biogeochemical cycles; however, controls on carbon and nitrogen ratios in aquatic environments are poorly understood. Here we use a multi-biome and global dataset (tropics to Arctic) of stream water chemistry to assess relationships between dissolved organic carbon (DOC) and nitrate, ammonium and dissolved organic nitrogen (DON), providing a new conceptual framework to consider interactions between DOC and the multiple forms of dissolved nitrogen. We found that across streams the total dissolved nitrogen (TDN) pool is comprised of very little ammonium and as DOC concentrations increase the TDN pool shifts from nitrate to DON dominated. This suggests that in high DOC systems, DON serves as the primary source of nitrogen. At the global scale, DOC and DON are positively correlated (r2 = 0.67) and the average C: N ratio of dissolved organic matter (molar ratio of DOC: DON) across our data set is approximately 31. At the biome and smaller regional scale the relationship between DOC and DON is highly variable (r2 = 0.07 - 0.56) with the strongest relationships found in streams draining the mixed temperate forests of the northeastern United States. DOC: DON relationships also display spatial and temporal variability including latitudinal and seasonal trends, and interactions with land-use. DOC: DON ratios correlated positively with gradients of energy versus nutrient limitation pointing to the ecological role (energy source versus nutrient source) that DON plays with stream ecosystems. Contrary to previous findings we found consistently weak relationships between DON and nitrate which may reflect DON's duality as an energy or nutrient source. Collectively these analyses demonstrate how gradients of DOC drive compositional changes in the TDN pool and reveal a high degree of variability in the C: N ratio

  16. Testing a community water supply well located near a stream for susceptibility to stream contamination and low-flows.

    Science.gov (United States)

    Stewart-Maddox, N. S.; Tysor, E. H.; Swanson, J.; Degon, A.; Howard, J.; Tsinnajinnie, L.; Frisbee, M. D.; Wilson, J. L.; Newman, B. D.

    2014-12-01

    A community well is the primary water supply to the town of El Rito. This small rural town in is located in a semi-arid, mountainous portion of northern New Mexico where water is scarce. The well is 72 meters from a nearby intermittent stream. Initial tritium sampling suggests a groundwater connection between the stream and well. The community is concerned with the sustainability and future quality of the well water. If this well is as tightly connected to the stream as the tritium data suggests, then the well is potentially at risk due to upstream contamination and the impacts of extended drought. To examine this, we observed the well over a two-week period performing pump and recovery tests, electrical resistivity surveys, and physical observations of the nearby stream. We also collected general chemistry, stable isotope and radon samples from the well and stream. Despite the large well diameter, our pump test data exhibited behavior similar to a Theis curve, but the rate of drawdown decreased below the Theis curve late in the test. This decrease suggests that the aquifer is being recharged, possibly through delayed yield, upwelling of groundwater, or from the stream. The delayed yield hypothesis is supported by our electrical resistivity surveys, which shows very little change in the saturated zone over the course of the pump test, and by low values of pump-test estimated aquifer storativity. Observations of the nearby stream showed no change in stream-water level throughout the pump test. Together this data suggests that the interaction between the stream and the well is low, but recharge could be occurring through other mechanisms such as delayed yield. Additional pump tests of longer duration are required to determine the exact nature of the aquifer and its communication with the well.

  17. Salting our landscape: An integrated catchment model using readily accessible data to assess emerging road salt contamination to streams

    International Nuclear Information System (INIS)

    Jin Li; Whitehead, Paul; Siegel, Donald I.; Findlay, Stuart

    2011-01-01

    A new integrated catchment model for salinity has been developed to assess the transport of road salt from upland areas in watersheds to streams using readily accessible landscape, hydrologic, and meteorological data together with reported salt applications. We used Fishkill Creek (NY) as a representative watershed to test the model. Results showed good agreement between modeled and measured stream water chloride concentrations. These results suggest that a dominant mode of catchment simulation that does not entail complex deterministic modeling is an appropriate method to model salinization and to assess effects of future applications of road salt to streams. We heuristically increased and decreased salt applications by 100% and results showed that stream chloride concentrations increased by 13% and decreased by 7%, respectively. The model suggests that future management of salt application can reduce environmental concentrations, albeit over some time. - Highlights: → A new Integrated Catchment Model (INCA-Cl) is developed to simulate salinity. → Road salt application is important in controlling stream chloride concentration. → INCA-Cl can be used to manage and forecast the input and transport of chloride to the rivers. - A newly developed integrated catchment model for salinity can be used to manage and forecast the inputs and transport of chloride to streams.

  18. Salting our landscape: An integrated catchment model using readily accessible data to assess emerging road salt contamination to streams

    Energy Technology Data Exchange (ETDEWEB)

    Jin Li, E-mail: li.jin@ouce.ox.ac.uk [Earth Sciences Department, Syracuse University, Syracuse, NY 13210 (United States); School of Geography and the Environment, University of Oxford, Oxford, OX1 3QY (United Kingdom); Whitehead, Paul [School of Geography and the Environment, University of Oxford, Oxford, OX1 3QY (United Kingdom); Siegel, Donald I. [Earth Sciences Department, Syracuse University, Syracuse, NY 13210 (United States); Findlay, Stuart [Cary Institute of Ecosystem Studies, 2801 Sharon Turnpike, Millbrook, NY 12545 (United States)

    2011-05-15

    A new integrated catchment model for salinity has been developed to assess the transport of road salt from upland areas in watersheds to streams using readily accessible landscape, hydrologic, and meteorological data together with reported salt applications. We used Fishkill Creek (NY) as a representative watershed to test the model. Results showed good agreement between modeled and measured stream water chloride concentrations. These results suggest that a dominant mode of catchment simulation that does not entail complex deterministic modeling is an appropriate method to model salinization and to assess effects of future applications of road salt to streams. We heuristically increased and decreased salt applications by 100% and results showed that stream chloride concentrations increased by 13% and decreased by 7%, respectively. The model suggests that future management of salt application can reduce environmental concentrations, albeit over some time. - Highlights: > A new Integrated Catchment Model (INCA-Cl) is developed to simulate salinity. > Road salt application is important in controlling stream chloride concentration. > INCA-Cl can be used to manage and forecast the input and transport of chloride to the rivers. - A newly developed integrated catchment model for salinity can be used to manage and forecast the inputs and transport of chloride to streams.

  19. Comparing two periphyton collection methods commonly used for stream bioassessment and the development of numeric nutrient standards.

    Science.gov (United States)

    Rodman, Ashley R; Scott, J Thad

    2017-07-01

    Periphyton is an important component of stream bioassessment, yet methods for quantifying periphyton biomass can differ substantially. A case study within the Arkansas Ozarks is presented to demonstrate the potential for linking chlorophyll-a (chl-a) and ash-free dry mass (AFDM) data sets amassed using two frequently used periphyton sampling protocols. Method A involved collecting periphyton from a known area on the top surface of variably sized rocks gathered from relatively swift-velocity riffles without discerning canopy cover. Method B involved collecting periphyton from the entire top surface of cobbles systematically gathered from riffle-run habitat where canopy cover was intentionally avoided. Chl-a and AFDM measurements were not different between methods (p = 0.123 and p = 0.550, respectively), and there was no interaction between method and time in the repeated measures structure of the study. However, significantly different seasonal distinctions were observed for chl-a and AFDM from all streams when data from the methods were combined (p methods may effectively be used together with some minor considerations due to potential confounding factors. This study provides motivation for the continued investigation of combining data sets derived from multiple methods of data collection, which could be useful in stream bioassessment and particularly important for the development of regional stream nutrient criteria for the southern Ozarks.

  20. ThermoData Engine (TDE): Software Implementation of the Dynamic Data Evaluation Concept. 8. Properties of Material Streams and Solvent Design

    DEFF Research Database (Denmark)

    Diky, Vladimir; Chirico, Robert D.; Muzny, Chris D.

    2013-01-01

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported in this journal. The present paper describes the first application of this concept to the evaluation of thermophysical properties for material streams involving any number...... of chemical components with assessment of uncertainties. The method involves construction of Redlich-Kister type equations for individual properties (excess volume, thermal conductivity, viscosity, surface tension, and excess enthalpy) and activity-coefficient models for phase equilibrium properties (vapor...

  1. Cost-efficient enactment of stream processing topologies

    Directory of Open Access Journals (Sweden)

    Christoph Hochreiner

    2017-12-01

    Full Text Available The continuous increase of unbound streaming data poses several challenges to established data stream processing engines. One of the most important challenges is the cost-efficient enactment of stream processing topologies under changing data volume. These data volume pose different loads to stream processing systems whose resource provisioning needs to be continuously updated at runtime. First approaches already allow for resource provisioning on the level of virtual machines (VMs, but this only allows for coarse resource provisioning strategies. Based on current advances and benefits for containerized software systems, we have designed a cost-efficient resource provisioning approach and integrated it into the runtime of the Vienna ecosystem for elastic stream processing. Our resource provisioning approach aims to maximize the resource usage for VMs obtained from cloud providers. This strategy only releases processing capabilities at the end of the VMs minimal leasing duration instead of releasing them eagerly as soon as possible as it is the case for threshold-based approaches. This strategy allows us to improve the service level agreement compliance by up to 25% and a reduction for the operational cost of up to 36%.

  2. The suitability of using dissolved gases to determine groundwater discharge to high gradient streams

    Science.gov (United States)

    Gleeson, Tom; Manning, Andrew H.; Popp, Andrea; Zane, Matthew; Clark, Jordan F.

    2018-02-01

    Determining groundwater discharge to streams using dissolved gases is known to be useful over a wide range of streamflow rates but the suitability of dissolved gas methods to determine discharge rates in high gradient mountain streams has not been sufficiently tested, even though headwater streams are critical as ecological habitats and water resources. The aim of this study is to test the suitability of using dissolved gases to determine groundwater discharge rates to high gradient streams by field experiments in a well-characterized, high gradient mountain stream and a literature review. At a reach scale (550 m) we combined stream and groundwater radon activity measurements with an in-stream SF6 tracer test. By means of numerical modeling we determined gas exchange velocities and derived very low groundwater discharge rates (∼15% of streamflow). These groundwater discharge rates are below the uncertainty range of physical streamflow measurements and consistent with temperature, specific conductance and streamflow measured at multiple locations along the reach. At a watershed-scale (4 km), we measured CFC-12 and δ18O concentrations and determined gas exchange velocities and groundwater discharge rates with the same numerical model. The groundwater discharge rates along the 4 km stream reach were highly variable, but were consistent with the values derived in the detailed study reach. Additionally, we synthesized literature values of gas exchange velocities for different stream gradients which show an empirical relationship that will be valuable in planning future dissolved gas studies on streams with various gradients. In sum, we show that multiple dissolved gas tracers can be used to determine groundwater discharge to high gradient mountain streams from reach to watershed scales.

  3. The suitability of using dissolved gases to determine groundwater discharge to high gradient streams

    Science.gov (United States)

    Gleeson, Tom; Manning, Andrew H.; Popp, Andrea; Zane, Mathew; Clark, Jordan F.

    2018-01-01

    Determining groundwater discharge to streams using dissolved gases is known to be useful over a wide range of streamflow rates but the suitability of dissolved gas methods to determine discharge rates in high gradient mountain streams has not been sufficiently tested, even though headwater streams are critical as ecological habitats and water resources. The aim of this study is to test the suitability of using dissolved gases to determine groundwater discharge rates to high gradient streams by field experiments in a well-characterized, high gradient mountain stream and a literature review. At a reach scale (550 m) we combined stream and groundwater radon activity measurements with an in-stream SF6 tracer test. By means of numerical modeling we determined gas exchange velocities and derived very low groundwater discharge rates (∼15% of streamflow). These groundwater discharge rates are below the uncertainty range of physical streamflow measurements and consistent with temperature, specific conductance and streamflow measured at multiple locations along the reach. At a watershed-scale (4 km), we measured CFC-12 and δ18O concentrations and determined gas exchange velocities and groundwater discharge rates with the same numerical model. The groundwater discharge rates along the 4 km stream reach were highly variable, but were consistent with the values derived in the detailed study reach. Additionally, we synthesized literature values of gas exchange velocities for different stream gradients which show an empirical relationship that will be valuable in planning future dissolved gas studies on streams with various gradients. In sum, we show that multiple dissolved gas tracers can be used to determine groundwater discharge to high gradient mountain streams from reach to watershed scales.

  4. Intensity Maps Production Using Real-Time Joint Streaming Data Processing From Social and Physical Sensors

    Science.gov (United States)

    Kropivnitskaya, Y. Y.; Tiampo, K. F.; Qin, J.; Bauer, M.

    2015-12-01

    Intensity is one of the most useful measures of earthquake hazard, as it quantifies the strength of shaking produced at a given distance from the epicenter. Today, there are several data sources that could be used to determine intensity level which can be divided into two main categories. The first category is represented by social data sources, in which the intensity values are collected by interviewing people who experienced the earthquake-induced shaking. In this case, specially developed questionnaires can be used in addition to personal observations published on social networks such as Twitter. These observations are assigned to the appropriate intensity level by correlating specific details and descriptions to the Modified Mercalli Scale. The second category of data sources is represented by observations from different physical sensors installed with the specific purpose of obtaining an instrumentally-derived intensity level. These are usually based on a regression of recorded peak acceleration and/or velocity amplitudes. This approach relates the recorded ground motions to the expected felt and damage distribution through empirical relationships. The goal of this work is to implement and evaluate streaming data processing separately and jointly from both social and physical sensors in order to produce near real-time intensity maps and compare and analyze their quality and evolution through 10-minute time intervals immediately following an earthquake. Results are shown for the case study of the M6.0 2014 South Napa, CA earthquake that occurred on August 24, 2014. The using of innovative streaming and pipelining computing paradigms through IBM InfoSphere Streams platform made it possible to read input data in real-time for low-latency computing of combined intensity level and production of combined intensity maps in near-real time. The results compare three types of intensity maps created based on physical, social and combined data sources. Here we correlate

  5. Stellar Streams Discovered in the Dark Energy Survey

    Energy Technology Data Exchange (ETDEWEB)

    Shipp, N.; et al.

    2018-01-09

    We perform a search for stellar streams around the Milky Way using the first three years of multi-band optical imaging data from the Dark Energy Survey (DES). We use DES data covering $\\sim 5000$ sq. deg. to a depth of $g > 23.5$ with a relative photometric calibration uncertainty of $< 1 \\%$. This data set yields unprecedented sensitivity to the stellar density field in the southern celestial hemisphere, enabling the detection of faint stellar streams to a heliocentric distance of $\\sim 50$ kpc. We search for stellar streams using a matched-filter in color-magnitude space derived from a synthetic isochrone of an old, metal-poor stellar population. Our detection technique recovers four previously known thin stellar streams: Phoenix, ATLAS, Tucana III, and a possible extension of Molonglo. In addition, we report the discovery of eleven new stellar streams. In general, the new streams detected by DES are fainter, more distant, and lower surface brightness than streams detected by similar techniques in previous photometric surveys. As a by-product of our stellar stream search, we find evidence for extra-tidal stellar structure associated with four globular clusters: NGC 288, NGC 1261, NGC 1851, and NGC 1904. The ever-growing sample of stellar streams will provide insight into the formation of the Galactic stellar halo, the Milky Way gravitational potential, as well as the large- and small-scale distribution of dark matter around the Milky Way.

  6. Optimization of the Darrieus wind turbines with double-multiple-streamtube model

    International Nuclear Information System (INIS)

    Paraschivoiu, I.

    1985-01-01

    This paper discusses a new improvement of the double-multiple-stream tube model by considering the stream tube expansion effects on the Darrieus wind turbine. These effects, allowing a more realistic modeling of the upwind/downwind flow field asymmetries inherent in the Darrieus rotor, were calculated by using CARDAAX computer code. When the dynamic stall is introduced in the double-multiple-stream tube model, the aerodynamic loads and performance show significant changes in the range of low tip-speed ratio

  7. Analytical solutions for the surface response to small amplitude perturbations in boundary data in the shallow-ice-stream approximation

    Directory of Open Access Journals (Sweden)

    G. H. Gudmundsson

    2008-07-01

    Full Text Available New analytical solutions describing the effects of small-amplitude perturbations in boundary data on flow in the shallow-ice-stream approximation are presented. These solutions are valid for a non-linear Weertman-type sliding law and for Newtonian ice rheology. Comparison is made with corresponding solutions of the shallow-ice-sheet approximation, and with solutions of the full Stokes equations. The shallow-ice-stream approximation is commonly used to describe large-scale ice stream flow over a weak bed, while the shallow-ice-sheet approximation forms the basis of most current large-scale ice sheet models. It is found that the shallow-ice-stream approximation overestimates the effects of bed topography perturbations on surface profile for wavelengths less than about 5 to 10 ice thicknesses, the exact number depending on values of surface slope and slip ratio. For high slip ratios, the shallow-ice-stream approximation gives a very simple description of the relationship between bed and surface topography, with the corresponding transfer amplitudes being close to unity for any given wavelength. The shallow-ice-stream estimates for the timescales that govern the transient response of ice streams to external perturbations are considerably more accurate than those based on the shallow-ice-sheet approximation. In particular, in contrast to the shallow-ice-sheet approximation, the shallow-ice-stream approximation correctly reproduces the short-wavelength limit of the kinematic phase speed given by solving a linearised version of the full Stokes system. In accordance with the full Stokes solutions, the shallow-ice-sheet approximation predicts surface fields to react weakly to spatial variations in basal slipperiness with wavelengths less than about 10 to 20 ice thicknesses.

  8. Generic accelerated sequence alignment in SeqAn using vectorization and multi-threading.

    Science.gov (United States)

    Rahn, René; Budach, Stefan; Costanza, Pascal; Ehrhardt, Marcel; Hancox, Jonny; Reinert, Knut

    2018-05-03

    Pairwise sequence alignment is undoubtedly a central tool in many bioinformatics analyses. In this paper, we present a generically accelerated module for pairwise sequence alignments applicable for a broad range of applications. In our module, we unified the standard dynamic programming kernel used for pairwise sequence alignments and extended it with a generalized inter-sequence vectorization layout, such that many alignments can be computed simultaneously by exploiting SIMD (Single Instruction Multiple Data) instructions of modern processors. We then extended the module by adding two layers of thread-level parallelization, where we a) distribute many independent alignments on multiple threads and b) inherently parallelize a single alignment computation using a work stealing approach producing a dynamic wavefront progressing along the minor diagonal. We evaluated our alignment vectorization and parallelization on different processors, including the newest Intel® Xeon® (Skylake) and Intel® Xeon Phi™ (KNL) processors, and use cases. The instruction set AVX512-BW (Byte and Word), available on Skylake processors, can genuinely improve the performance of vectorized alignments. We could run single alignments 1600 times faster on the Xeon Phi™ and 1400 times faster on the Xeon® than executing them with our previous sequential alignment module. The module is programmed in C++ using the SeqAn (Reinert et al., 2017) library and distributed with version 2.4. under the BSD license. We support SSE4, AVX2, AVX512 instructions and included UME::SIMD, a SIMD-instruction wrapper library, to extend our module for further instruction sets. We thoroughly test all alignment components with all major C++ compilers on various platforms. rene.rahn@fu-berlin.de.

  9. Supervised classification of distributed data streams for smart grids

    Energy Technology Data Exchange (ETDEWEB)

    Guarracino, Mario R. [High Performance Computing and Networking - National Research Council of Italy, Naples (Italy); Irpino, Antonio; Verde, Rosanna [Seconda Universita degli Studi di Napoli, Dipartimento di Studi Europei e Mediterranei, Caserta (Italy); Radziukyniene, Neringa [Lithuanian Energy Institute, Laboratory of Systems Control and Automation, Kaunas (Lithuania)

    2012-03-15

    The electricity system inherited from the 19th and 20th centuries has been a reliable but centralized system. With the spreading of local, distributed and intermittent renewable energy resources, top-down central control of the grid no longer meets modern requirements. For these reasons, the power grid has been equipped with smart meters integrating bi-directional communications, advanced power measurement and management capabilities. Smart meters make it possible to remotely turn power on or off to a customer, read usage information, detect a service outage and the unauthorized use of electricity. To fully exploit their capabilities, we foresee the usage of distributed supervised classification algorithms. By gathering data available from meters and other sensors, such algorithms can create local classification models for attack detection, online monitoring, privacy preservation, workload balancing, prediction of energy demand and incoming faults. In this paper we present a decentralized distributed classification algorithm based on proximal support vector machines. The method uses partial knowledge, in form of data streams, to build its local model on each meter. We demonstrate the performance of the proposed scheme on synthetic datasets. (orig.)

  10. Organic carbon spiralling in stream ecosystems

    Energy Technology Data Exchange (ETDEWEB)

    Newbold, J D; Mulholland, P J; Elwood, J W; O' Neill, R V

    1982-01-01

    The term spiralling has been used to describe the combined processes of cycling and longitudinal transport in streams. As a measure or organic carbon spiralling, we introduced organic carbon turnover length, S, defined as the average or expected downstream distance travelled by a carbon atom between its entry or fixation in the stream and its oxidation. Using a simple model for organic carbon dynamics in a stream, we show that S is closely related to fisher and Likens' ecosystem efficiency. Unlike efficiency, however, S is independent of the length of the study reach, and values of S determined in streams of differing lengths can be compared. Using data from three different streams, we found the relationship between S and efficiency to agree closely with the model prediction. Hypotheses of stream functioning are discussed in the context of organic carbeon spiralling theory.

  11. A Linear Algebra Framework for Static High Performance Fortran Code Distribution

    Directory of Open Access Journals (Sweden)

    Corinne Ancourt

    1997-01-01

    Full Text Available High Performance Fortran (HPF was developed to support data parallel programming for single-instruction multiple-data (SIMD and multiple-instruction multiple-data (MIMD machines with distributed memory. The programmer is provided a familiar uniform logical address space and specifies the data distribution by directives. The compiler then exploits these directives to allocate arrays in the local memories, to assign computations to elementary processors, and to migrate data between processors when required. We show here that linear algebra is a powerful framework to encode HPF directives and to synthesize distributed code with space-efficient array allocation, tight loop bounds, and vectorized communications for INDEPENDENT loops. The generated code includes traditional optimizations such as guard elimination, message vectorization and aggregation, and overlap analysis. The systematic use of an affine framework makes it possible to prove the compilation scheme correct.

  12. Organism-substrate relationships in lowland streams

    NARCIS (Netherlands)

    Tolkamp, H.H.

    1980-01-01

    A field and laboratory study on the microdistribution of bottom dwelling macroinvertebrates to investigate the role of the stream substrate In the development and preservation of the macroinvertebrate communities in natural, undisturbed lowland streams is described. Field data on bottom substrates

  13. Minería de datos sobre streams de redes sociales, una herramienta al servicio de la Bibliotecología = Data Mining Streams of Social Networks, A Tool to Improve The Library Services

    Directory of Open Access Journals (Sweden)

    Sonia Jaramillo Valbuena

    2015-12-01

    , Facebook, RSS feeds and blogs, generate a large amount of unstructured data streams. They can be used to the problem of mining topic-specific influence, graph mining, opinion mining and recommender systems, thus achieving that libraries can obtain maximum benefit from the use of Information and Communication Technologies. From the perspective of data stream mining, the processing of these streams poses significant challenges. The algorithms must be adapted to problems such as: high arrival rate, memory requirements without restrictions, diverse sources of data and concept-drift. In this work, we explore the current state-of-the-art solutions of data stream mining originating from social networks, specifically, Facebook and Twitter. We present a review of the most representative algorithms and how they contribute to knowledge discovery in the area of librarianship. We conclude by presenting some of the problems that are the subject of active research.

  14. Sources and preparation of data for assessing trends in concentrations of pesticides in streams of the United States, 1992–2010

    Science.gov (United States)

    Martin, Jeffrey D.; Eberle, Michael; Nakagaki, Naomi

    2011-01-01

    This report updates a previously published water-quality dataset of 44 commonly used pesticides and 8 pesticide degradates suitable for a national assessment of trends in pesticide concentrations in streams of the United States. Water-quality samples collected from January 1992 through September 2010 at stream-water sites of the U.S. Geological Survey (USGS) National Water-Quality Assessment (NAWQA) Program and the National Stream Quality Accounting Network (NASQAN) were compiled, reviewed, selected, and prepared for trend analysis. The principal steps in data review for trend analysis were to (1) identify analytical schedule, (2) verify sample-level coding, (3) exclude inappropriate samples or results, (4) review pesticide detections per sample, (5) review high pesticide concentrations, and (6) review the spatial and temporal extent of NAWQA pesticide data and selection of analytical methods for trend analysis. The principal steps in data preparation for trend analysis were to (1) select stream-water sites for trend analysis, (2) round concentrations to a consistent level of precision for the concentration range, (3) identify routine reporting levels used to report nondetections unaffected by matrix interference, (4) reassign the concentration value for routine nondetections to the maximum value of the long-term method detection level (maxLT-MDL), (5) adjust concentrations to compensate for temporal changes in bias of recovery of the gas chromatography/mass spectrometry (GCMS) analytical method, and (6) identify samples considered inappropriate for trend analysis. Samples analyzed at the USGS National Water Quality Laboratory (NWQL) by the GCMS analytical method were the most extensive in time and space and, consequently, were selected for trend analysis. Stream-water sites with 3 or more water years of data with six or more samples per year were selected for pesticide trend analysis. The selection criteria described in the report produced a dataset of 21

  15. Urban development results in stressors that degrade stream ecosystems

    Science.gov (United States)

    Bell, Amanda H.; Coles, James F.; McMahon, Gerard; Woodside, Michael D.

    2012-01-01

    In 2003, eighty-three percent of Americans lived in metropolitan areas, and considerable population increases are predicted within the next 50 years. Nowhere are the environmental changes associated with urban development more evident than in urban streams. Contaminants, habitat destruction, and increasing streamflow flashiness resulting from urban development have been associated with the disruption of biological communities, particularly the loss of sensitive aquatic biota. Every stream is connected downstream to other water bodies, and inputs of contaminants and (or) sediments to streams can cause degradation downstream with adverse effects on biological communities and on economically valuable resources, such as fisheries and tourism. Understanding how algal, invertebrate, and fish communities respond to physical and chemical stressors associated with urban development can provide important clues on how multiple stressors may be managed to protect stream health as a watershed becomes increasingly urbanized. This fact sheet highlights selected findings of a comprehensive assessment by the National Water-Quality Assessment Program of the U.S. Geological Survey (USGS) of the effects of urban development on stream ecosystems in nine metropolitan study areas.

  16. MP CBM-Z V1.0: design for a new CBM-Z gas-phase chemical mechanism architecture for next generation processors

    OpenAIRE

    Wang, Hui; Lin, Junmin; Wu, Qizhong; Chen, Huansheng; Tang, Xiao; Wang, Zifa; Chen, Xueshun; Cheng, Huaqiong; Wang, Lanning

    2018-01-01

    Precise and rapid air quality simulation and forecasting are limited by the computation performance of the air quality model, and the gas-phase chemistry module is the most time-consuming function in the air quality model. In this study, we designed a new framework for the widely used Carbon Bond Mechanism Z (CBM-Z) gas-phase chemical kinetics kernel to adapt the Single Instruction Multiple Data (SIMD) technology in the next-generation processors for improving its calculation performance. The...

  17. Hydrogeochemical and stream sediment reconnaissance basic data for Brownsville-McAllen NTMS Quadrangles, Texas

    International Nuclear Information System (INIS)

    1980-01-01

    Results of a reconnaissance geochemical survey of the Brownsville-McAllen Quadrangles, Texas are reported. Field and laboratory data are presented for 427 groundwater and 171 stream sediment samples. Statistical and areal distributions of uranium and possible uranium-related variables are displayed. Pertinent geologic factors which may be of significance in evaluating the potential for uranium mineralization are briefly discussed. Groundwater data indicate the most promising area for potential uranium mineralization occurs in the northwestern section of the quadrangles (Jim Hogg, Starr, and Zapata Counties), where waters are derived from the Catahoula Formation. These groundwaters have high concentrations of uranium, uranium associated elements, and low values for specific conductance. Another area with high uranium concentrations is in the southeastern portion of the survey area (Hidalgo, Cameron, and Willacy Counties). Shallow wells <10 m (30 ft) are numerous in this area and high specific conductance values may indicate contamination from extensive fertilization. Stream sediment data for the survey does not indicate an area favorable for uranium mineralization. Anomalous acid soluble uranium values in the southeastern area (Hidalgo, Cameron, and Willacy Counties) can be attributed to phosphate fertilizer contamination. Four samples in the western part of the area (western Starr County) have anomalously high total uranium values and low acid soluble uranium values, indicating the uranium may be contained in resistate minerals

  18. Two-Tier VoI Prioritization System on Requirement-Based Data Streaming toward IoT

    Directory of Open Access Journals (Sweden)

    Sunyanan Choochotkaew

    2017-01-01

    Full Text Available Toward the world of Internet of Things, people utilize knowledge from sensor streams in various kinds of smart applications. The number of sensing devices is rapidly increasing along with the amount of sensing data. Consequently, a bottleneck problem at the local gateway has attracted high concern. An example scenario is smart elderly houses in rural areas where each house installs thousands of sensors and all connect to resource-limited and unstable 2G/3G networks. The bottleneck state can incur unacceptable latency and loss of significant data due to the limited waiting-queue. Orthogonally to the existing solutions, we propose a two-tier prioritization system to enhance information quality, indicated by VoI, at the local gateway. The proposed system has been designed to support several requirements with several conflicting criteria over shared sensing streams. Our approach adopts Multicriteria Decision Analysis technique to merge requirements and to assess the VoI. We introduce the framework that can reduce the computational cost by precalculation. Through a case study of building management systems, we have shown that our merge algorithm can provide 0.995 cosine-similarity for representing all user requirements and the evaluation approach can obtain satisfaction values around 3 times higher than the naïve strategies for the top-list data.

  19. Streaming simplification of tetrahedral meshes.

    Science.gov (United States)

    Vo, Huy T; Callahan, Steven P; Lindstrom, Peter; Pascucci, Valerio; Silva, Cláudio T

    2007-01-01

    Unstructured tetrahedral meshes are commonly used in scientific computing to represent scalar, vector, and tensor fields in three dimensions. Visualization of these meshes can be difficult to perform interactively due to their size and complexity. By reducing the size of the data, we can accomplish real-time visualization necessary for scientific analysis. We propose a two-step approach for streaming simplification of large tetrahedral meshes. Our algorithm arranges the data on disk in a streaming, I/O-efficient format that allows coherent access to the tetrahedral cells. A quadric-based simplification is sequentially performed on small portions of the mesh in-core. Our output is a coherent streaming mesh which facilitates future processing. Our technique is fast, produces high quality approximations, and operates out-of-core to process meshes too large for main memory.

  20. Submersible Data (Dive Waypoints) for Islands in the Stream 2002 - Deep Reef Habitat - Office of Ocean Exploration

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Data and information collected by the submersible Johnson Sea-Link II at waypoints along its track during one dive of the 2002 "Islands in the Stream - Deep Reef...

  1. Programmable stream prefetch with resource optimization

    Science.gov (United States)

    Boyle, Peter; Christ, Norman; Gara, Alan; Mawhinney, Robert; Ohmacht, Martin; Sugavanam, Krishnan

    2013-01-08

    A stream prefetch engine performs data retrieval in a parallel computing system. The engine receives a load request from at least one processor. The engine evaluates whether a first memory address requested in the load request is present and valid in a table. The engine checks whether there exists valid data corresponding to the first memory address in an array if the first memory address is present and valid in the table. The engine increments a prefetching depth of a first stream that the first memory address belongs to and fetching a cache line associated with the first memory address from the at least one cache memory device if there is not yet valid data corresponding to the first memory address in the array. The engine determines whether prefetching of additional data is needed for the first stream within its prefetching depth. The engine prefetches the additional data if the prefetching is needed.

  2. The effects of road crossings on prairie stream habitat and function

    Science.gov (United States)

    Bouska, Wesley W.; Keane, Timothy; Paukert, Craig P.

    2010-01-01

    Improperly designed stream crossing structures may alter the form and function of stream ecosystems and habitat and prohibit the movement of aquatic organisms. Stream sections adjoining five concrete box culverts, five low-water crossings (concrete slabs vented by one or multiple culverts), and two large, single corrugated culvert vehicle crossings in eastern Kansas streams were compared to reference reaches using a geomorphologic survey and stream classification. Stream reaches were also compared upstream and downstream of crossings, and crossing measurements were used to determine which crossing design best mimicked the natural dimensions of the adjoining stream. Four of five low-water crossings, three of five box culverts, and one of two large, single corrugated pipe culverts changed classification from upstream to downstream of the crossings. Mean riffle spacing upstream at low-water crossings (8.6 bankfull widths) was double that of downstream reaches (mean 4.4 bankfull widths) but was similar upstream and downstream of box and corrugated pipe culverts. There also appeared to be greater deposition of fine sediments directly upstream of these designs. Box and corrugated culverts were more similar to natural streams than low-water crossings at transporting water, sediments, and debris during bankfull flows.

  3. Noise Prediction Module for Offset Stream Nozzles

    Science.gov (United States)

    Henderson, Brenda S.

    2011-01-01

    A Modern Design of Experiments (MDOE) analysis of data acquired for an offset stream technology was presented. The data acquisition and concept development were funded under a Supersonics NRA NNX07AC62A awarded to Dimitri Papamoschou at University of California, Irvine. The technology involved the introduction of airfoils in the fan stream of a bypass ratio (BPR) two nozzle system operated at transonic exhaust speeds. The vanes deflected the fan stream relative to the core stream and resulted in reduced sideline noise for polar angles in the peak jet noise direction. Noise prediction models were developed for a range of vane configurations. The models interface with an existing ANOPP module and can be used or future system level studies.

  4. Effects of soil data resolution on SWAT model stream flow and water quality predictions.

    Science.gov (United States)

    Geza, Mengistu; McCray, John E

    2008-08-01

    The prediction accuracy of agricultural nonpoint source pollution models such as Soil and Water Assessment Tool (SWAT) depends on how well model input spatial parameters describe the characteristics of the watershed. The objective of this study was to assess the effects of different soil data resolutions on stream flow, sediment and nutrient predictions when used as input for SWAT. SWAT model predictions were compared for the two US Department of Agriculture soil databases with different resolution, namely the State Soil Geographic database (STATSGO) and the Soil Survey Geographic database (SSURGO). Same number of sub-basins was used in the watershed delineation. However, the number of HRUs generated when STATSGO and SSURGO soil data were used is 261 and 1301, respectively. SSURGO, with the highest spatial resolution, has 51 unique soil types in the watershed distributed in 1301 HRUs, while STATSGO has only three distributed in 261 HRUS. As a result of low resolution STATSGO assigns a single classification to areas that may have different soil types if SSURGO were used. SSURGO included Hydrologic Response Units (HRUs) with soil types that were generalized to one soil group in STATSGO. The difference in the number and size of HRUs also has an effect on sediment yield parameters (slope and slope length). Thus, as a result of the discrepancies in soil type and size of HRUs stream flow predicted was higher when SSURGO was used compared to STATSGO. SSURGO predicted less stream loading than STATSGO in terms of sediment and sediment-attached nutrients components, and vice versa for dissolved nutrients. When compared to mean daily measured flow, STATSGO performed better relative to SSURGO before calibration. SSURGO provided better results after calibration as evaluated by R(2) value (0.74 compared to 0.61 for STATSGO) and the Nash-Sutcliffe coefficient of Efficiency (NSE) values (0.70 and 0.61 for SSURGO and STATSGO, respectively) although both are in the same satisfactory

  5. Roanoke 10 x 20 NTMS area, Virginia. Data report (abbreviated): National Uranium Resource Evaluation program, hydrogeochemical and stream sediment reconnaissance

    International Nuclear Information System (INIS)

    Cook, J.R.

    1980-12-01

    This abbreviated data report presents results of ground water and stream sediment reconnaissance in the National Topographic Map Series Roanoke 1 0 x 2 0 quadrangle. Surface sediment samples were collected at 1235 sites. Ground water samples were collected at 767 sites. Neutron activation analysis results are given for uranium and 16 other elements in sediments, and for uranium and 8 other elements in ground water. Field measurements and observations are reported for each site. Analytical data and field measurements are presented. Data from ground water sites include (1) water chemistry measurements (pH, conductivity, and alkalinity), (2) physical measurements where applicable (water temperature, well description, etc.), and (3) elemental analyses (U, Al, Br, Cl, Dy, F, Mn, Na, and V). Data from sediment sites include (1) stream water chemistry measurements (pH, conductivity, and alkalinity), and (2) elemental analyses for sediment samples (U, Th, Hf, Al, Ce, Dy, Eu, Fe, La, Lu, Mn, Sc, Sm, Na, Ti, V, and Yb). Sample site descriptors (stream characteristics, vegetation, etc.) are tabulated. Areal distribution maps, histograms, and cumulative frequency plots for most elements and for U/Th and U/Hf ratios are included. Key data from stream water sites include (1) water quality measurements (pH, conductivity, and alkalinity) and (2) elemental analyses (U, Al, Br, Cl, Dy, F, Mg, Mn, Na, and V). Uranium concentrations in the sediments range from 0.50 to 83.50 ppM with a mean of 6.67 ppM. A cluster of high log (U/Th + Hf) ratios appear in the southeastern portion of the quadrangle. Uranium, thorium, and the rare earth elements show a striking correlation with the geology of the area

  6. Sources of trends in water-quality data for selected streams in Texas, 1975-89 water years

    Science.gov (United States)

    Schertz, T.L.; Wells, F.C.; Ohe, D.J.

    1994-01-01

    Sources of trends in water-quality data for selected streams in Texas for the 1975-89 water years were investigated in this study. The investigation of sources was confined to distinct geographic patterns in the trend indicators for one constituent or for a group of related constituents.

  7. A Data Stream Model For Runoff Simulation In A Changing Environment

    Science.gov (United States)

    Yang, Q.; Shao, J.; Zhang, H.; Wang, G.

    2017-12-01

    Runoff simulation is of great significance for water engineering design, water disaster control, water resources planning and management in a catchment or region. A large number of methods including concept-based process-driven models and statistic-based data-driven models, have been proposed and widely used in worldwide during past decades. Most existing models assume that the relationship among runoff and its impacting factors is stationary. However, in the changing environment (e.g., climate change, human disturbance), their relationship usually evolves over time. In this study, we propose a data stream model for runoff simulation in a changing environment. Specifically, the proposed model works in three steps: learning a rule set, expansion of a rule, and simulation. The first step is to initialize a rule set. When a new observation arrives, the model will check which rule covers it and then use the rule for simulation. Meanwhile, Page-Hinckley (PH) change detection test is used to monitor the online simulation error of each rule. If a change is detected, the corresponding rule is removed from the rule set. In the second step, for each rule, if it covers more than a given number of instance, the rule is expected to expand. In the third step, a simulation model of each leaf node is learnt with a perceptron without activation function, and is updated with adding a newly incoming observation. Taking Fuxi River catchment as a case study, we applied the model to simulate the monthly runoff in the catchment. Results show that abrupt change is detected in the year of 1997 by using the Page-Hinckley change detection test method, which is consistent with the historic record of flooding. In addition, the model achieves good simulation results with the RMSE of 13.326, and outperforms many established methods. The findings demonstrated that the proposed data stream model provides a promising way to simulate runoff in a changing environment.

  8. Stream hydraulics and temperature determine the metabolism of geothermal Icelandic streams

    Directory of Open Access Journals (Sweden)

    Demars B. O.L.

    2011-07-01

    Full Text Available Stream ecosystem metabolism plays a critical role in planetary biogeochemical cycling. Stream benthic habitat complexity and the available surface area for microbes relative to the free-flowing water volume are thought to be important determinants of ecosystem metabolism. Unfortunately, the engineered deepening and straightening of streams for drainage purposes could compromise stream natural services. Stream channel complexity may be quantitatively expressed with hydraulic parameters such as water transient storage, storage residence time, and water spiralling length. The temperature dependence of whole stream ecosystem respiration (ER, gross primary productivity (GPP and net ecosystem production (NEP = GPP − ER has recently been evaluated with a “natural experiment” in Icelandic geothermal streams along a 5–25 °C temperature gradient. There remained, however, a substantial amount of unexplained variability in the statistical models, which may be explained by hydraulic parameters found to be unrelated to temperature. We also specifically tested the additional and predicted synergistic effects of water transient storage and temperature on ER, using novel, more accurate, methods. Both ER and GPP were highly related to water transient storage (or water spiralling length but not to the storage residence time. While there was an additional effect of water transient storage and temperature on ER (r2 = 0.57; P = 0.015, GPP was more related to water transient storage than temperature. The predicted synergistic effect could not be confirmed, most likely due to data limitation. Our interpretation, based on causal statistical modelling, is that the metabolic balance of streams (NEP was primarily determined by the temperature dependence of respiration. Further field and experimental work is required to test the predicted synergistic effect on ER. Meanwhile, since higher metabolic activities allow for higher pollutant degradation or uptake

  9. BLOSTREAM: A HIGH SPEED STREAM CIPHER

    Directory of Open Access Journals (Sweden)

    ALI H. KASHMAR

    2017-04-01

    Full Text Available Although stream ciphers are widely utilized to encrypt sensitive data at fast speeds, security concerns have led to a shift from stream to block ciphers, judging that the current technology in stream cipher is inferior to the technology of block ciphers. This paper presents the design of an improved efficient and secure stream cipher called Blostream, which is more secure than conventional stream ciphers that use XOR for mixing. The proposed cipher comprises two major components: the Pseudo Random Number Generator (PRNG using the Rabbit algorithm and a nonlinear invertible round function (combiner for encryption and decryption. We evaluate its performance in terms of implementation and security, presenting advantages and disadvantages, comparison of the proposed cipher with similar systems and a statistical test for randomness. The analysis shows that the proposed cipher is more efficient, high speed, and secure than current conventional stream ciphers.

  10. Monitoring stream temperatures—A guide for non-specialists

    Science.gov (United States)

    Heck, Michael P.; Schultz, Luke D.; Hockman-Wert, David; Dinger, Eric C.; Dunham, Jason B.

    2018-04-19

    Executive SummaryWater temperature influences most physical and biological processes in streams, and along with streamflows is a major driver of ecosystem processes. Collecting data to measure water temperature is therefore imperative, and relatively straightforward. Several protocols exist for collecting stream temperature data, but these are frequently directed towards specialists. This document was developed to address the need for a protocol intended for non-specialists (non-aquatic) staff. It provides specific step-by-step procedures on (1) how to launch data loggers, (2) check the factory calibration of data loggers prior to field use, (3) how to install data loggers in streams for year-round monitoring, (4) how to download and retrieve data loggers from the field, and (5) how to input project data into organizational databases.

  11. Ecoregions and stream morphology in eastern Oklahoma

    Science.gov (United States)

    Splinter, D.K.; Dauwalter, D.C.; Marston, R.A.; Fisher, W.L.

    2010-01-01

    Broad-scale variables (i.e., geology, topography, climate, land use, vegetation, and soils) influence channel morphology. How and to what extent the longitudinal pattern of channel morphology is influenced by broad-scale variables is important to fluvial geomorphologists and stream ecologists. In the last couple of decades, there has been an increase in the amount of interdisciplinary research between fluvial geomorphologists and stream ecologists. In a historical context, fluvial geomorphologists are more apt to use physiographic regions to distinguish broad-scale variables, while stream ecologists are more apt to use the concept of an ecosystem to address the broad-scale variables that influence stream habitat. For this reason, we designed a study using ecoregions, which uses physical and biological variables to understand how landscapes influence channel processes. Ecoregions are delineated by similarities in geology, climate, soils, land use, and potential natural vegetation. In the fluvial system, stream form and function are dictated by processes observed throughout the fluvial hierarchy. Recognizing that stream form and function should differ by ecoregion, a study was designed to evaluate how the characteristics of stream channels differed longitudinally among three ecoregions in eastern Oklahoma, USA: Boston Mountains, Ozark Highlands, and Ouachita Mountains. Channel morphology of 149 stream reaches was surveyed in 1st- through 4th-order streams, and effects of drainage area and ecoregion on channel morphology was evaluated using multiple regressions. Differences existed (?????0.05) among ecoregions for particle size, bankfull width, and width/depth ratio. No differences existed among ecoregions for gradient or sinuosity. Particle size was smallest in the Ozark Highlands and largest in the Ouachita Mountains. Bankfull width was larger in the Ozark Highlands than in the Boston Mountains and Ouachita Mountains in larger streams. Width/depth ratios of the

  12. Design and methods of the Pacific Northwest Stream Quality Assessment (PNSQA), 2015

    Science.gov (United States)

    Sheibley, Rich W.; Morace, Jennifer L.; Journey, Celeste A.; Van Metre, Peter C.; Bell, Amanda H.; Nakagaki, Naomi; Button, Daniel T.; Qi, Sharon L.

    2017-08-25

    In 2015, the U.S. Geological Survey (USGS) National Water-Quality Assessment (NAWQA) project conducted the Pacific Northwest Stream Quality Assessment (PNSQA) to investigate stream quality across the western part of the Pacific Northwest. The goal of the PNSQA was to assess the health of streams in the region by characterizing multiple water-quality factors that are stressors to in-stream aquatic life and by evaluating the relation between these stressors and the condition of biological communities. The effects of urbanization and agriculture on stream quality for the Puget Lowland and Willamette Valley Level III Ecoregions were the focus of this regional study. Findings will help inform the public and policymakers about human and environmental factors that are the most critical in affecting stream quality and, thus, provide insights into possible strategies to protect or improve the health of streams in the region.Land-use data were used in the study to identify and select sites within the region that ranged in levels of urban and agricultural development. A total of 88 sites were selected across the region—69 were on streams that explicitly spanned a range of urban land use in their watersheds, 8 were on streams in agricultural watersheds, and 11 were reference sites with little or no development in their watersheds. Depending on the type of land use, sites were sampled for contaminants, nutrients, and sediment for either a 4- or 10-week period during April, May, and June 2015. This water-quality “index period” was immediately followed with an ecological survey of all sites that included stream habitat, benthic algae, benthic macroinvertebrates, and fish. Additionally, streambed sediment was collected during the ecological survey for analysis of sediment chemistry and toxicity testing.This report provides a detailed description of the specific study components and methods of the PNSQA, including (1) surveys of stream habitat and aquatic biota, (2) discrete

  13. Assessing the chemical contamination dynamics in a mixed land use stream system

    DEFF Research Database (Denmark)

    Sonne, Anne Thobo; McKnight, Ursula S.; Rønde, Vinni

    2017-01-01

    Traditionally, the monitoring of streams for chemical and ecological status has been limited to surface water concentrations, where the dominant focus has been on general water quality and the risk for eutrophication. Mixed land use stream systems, comprising urban areas and agricultural production......, are challenging to assess with multiple chemical stressors impacting stream corridors. New approaches are urgently needed for identifying relevant sources, pathways and potential impacts for implementation of suitable source management and remedial measures. We developed a method for risk assessing chemical...... stressors in these systems and applied the approach to a 16-km groundwater-fed stream corridor (Grindsted, Denmark). Three methods were combined: (i) in-stream contaminant mass discharge for source quantification, (ii) Toxic Units and (iii) environmental standards. An evaluation of the chemical quality...

  14. Estimating autotrophic respiration in streams using daily metabolism data

    Science.gov (United States)

    Knowing the fraction of gross primary production (GPP) that is immediately respired by autotrophs and their closely associated heterotrophs (ARf) is necessary to understand the trophic base and carbon spiraling in streams. We show a means to estimate ARf from daily metabolism da...

  15. Ebullitive methane emissions from oxygenated wetland streams

    Science.gov (United States)

    Crawford, John T.; Stanley, Emily H.; Spawn, Seth A.; Finlay, Jacques C.; Striegl, Robert G.

    2014-01-01

    Stream and river carbon dioxide emissions are an important component of the global carbon cycle. Methane emissions from streams could also contribute to regional or global greenhouse gas cycling, but there are relatively few data regarding stream and river methane emissions. Furthermore, the available data do not typically include the ebullitive (bubble-mediated) pathway, instead focusing on emission of dissolved methane by diffusion or convection. Here, we show the importance of ebullitive methane emissions from small streams in the regional greenhouse gas balance of a lake and wetland-dominated landscape in temperate North America and identify the origin of the methane emitted from these well-oxygenated streams. Stream methane flux densities from this landscape tended to exceed those of nearby wetland diffusive fluxes as well as average global wetland ebullitive fluxes. Total stream ebullitive methane flux at the regional scale (103 Mg C yr−1; over 6400 km2) was of the same magnitude as diffusive methane flux previously documented at the same scale. Organic-rich stream sediments had the highest rates of bubble release and higher enrichment of methane in bubbles, but glacial sand sediments also exhibited high bubble emissions relative to other studied environments. Our results from a database of groundwater chemistry support the hypothesis that methane in bubbles is produced in anoxic near-stream sediment porewaters, and not in deeper, oxygenated groundwaters. Methane interacts with other key elemental cycles such as nitrogen, oxygen, and sulfur, which has implications for ecosystem changes such as drought and increased nutrient loading. Our results support the contention that streams, particularly those draining wetland landscapes of the northern hemisphere, are an important component of the global methane cycle.

  16. Integrated assessment of chemical stressors and ecological impact in mixed land use stream systems

    DEFF Research Database (Denmark)

    Sonne, Anne Thobo

    activities, including contaminated sites. To determine potential impacts, the chemical quality of both organic (i.e. pharmaceuticals, gasoline constituents, chlorinated solvents, and pesticides) and inorganic (i.e. metals, general water chemistry and macroions) compounds was assessed in all three stream...... multiple compounds (i.e. organic and inorganic chemical stressors) and stream compartments to locate key sources and risk drivers. The approaches and findings in this thesis could truly be helpful for management and future remediation of mixed land use stream systems....... of the different stream compartments thus comprises both temporal and spatial variation. Despite the growing understanding of the complexity, approaches for a holistic risk assessment of the potential impacts in the three stream compartments of a mixed land use stream system are still missing. To investigate...

  17. How General-Purpose can a GPU be?

    Directory of Open Access Journals (Sweden)

    Philip Machanick

    2015-12-01

    Full Text Available The use of graphics processing units (GPUs in general-purpose computation (GPGPU is a growing field. GPU instruction sets, while implementing a graphics pipeline, draw from a range of single instruction multiple datastream (SIMD architectures characteristic of the heyday of supercomputers. Yet only one of these SIMD instruction sets has been of application on a wide enough range of problems to survive the era when the full range of supercomputer design variants was being explored: vector instructions. This paper proposes a reconceptualization of the GPU as a multicore design with minimal exotic modes of parallelism so as to make GPGPU truly general.

  18. Hydrogeochemical and stream sediment reconnaissance basic data for Milbank NTMS Quadrangle, Minnesota; North Dakota; South Dakota

    International Nuclear Information System (INIS)

    1981-01-01

    Results of a reconnaissance geochemical survey are reported for the Milbank Quadrangle, Minnesota; North Dakota; South Dakota. Statistical data and areal distributions for uranium and uranium-related variables are presented for 662 groundwater and 319 stream sediment samples. Also included is a brief discussion on location and geologic setting

  19. Retrieval of Sentence Sequences for an Image Stream via Coherence Recurrent Convolutional Networks.

    Science.gov (United States)

    Park, Cesc Chunseong; Kim, Youngjin; Kim, Gunhee

    2018-04-01

    We propose an approach for retrieving a sequence of natural sentences for an image stream. Since general users often take a series of pictures on their experiences, much online visual information exists in the form of image streams, for which it would better take into consideration of the whole image stream to produce natural language descriptions. While almost all previous studies have dealt with the relation between a single image and a single natural sentence, our work extends both input and output dimension to a sequence of images and a sequence of sentences. For retrieving a coherent flow of multiple sentences for a photo stream, we propose a multimodal neural architecture called coherence recurrent convolutional network (CRCN), which consists of convolutional neural networks, bidirectional long short-term memory (LSTM) networks, and an entity-based local coherence model. Our approach directly learns from vast user-generated resource of blog posts as text-image parallel training data. We collect more than 22 K unique blog posts with 170 K associated images for the travel topics of NYC, Disneyland , Australia, and Hawaii. We demonstrate that our approach outperforms other state-of-the-art image captioning methods for text sequence generation, using both quantitative measures and user studies via Amazon Mechanical Turk.

  20. PRESEE: an MDL/MML algorithm to time-series stream segmenting.

    Science.gov (United States)

    Xu, Kaikuo; Jiang, Yexi; Tang, Mingjie; Yuan, Changan; Tang, Changjie

    2013-01-01

    Time-series stream is one of the most common data types in data mining field. It is prevalent in fields such as stock market, ecology, and medical care. Segmentation is a key step to accelerate the processing speed of time-series stream mining. Previous algorithms for segmenting mainly focused on the issue of ameliorating precision instead of paying much attention to the efficiency. Moreover, the performance of these algorithms depends heavily on parameters, which are hard for the users to set. In this paper, we propose PRESEE (parameter-free, real-time, and scalable time-series stream segmenting algorithm), which greatly improves the efficiency of time-series stream segmenting. PRESEE is based on both MDL (minimum description length) and MML (minimum message length) methods, which could segment the data automatically. To evaluate the performance of PRESEE, we conduct several experiments on time-series streams of different types and compare it with the state-of-art algorithm. The empirical results show that PRESEE is very efficient for real-time stream datasets by improving segmenting speed nearly ten times. The novelty of this algorithm is further demonstrated by the application of PRESEE in segmenting real-time stream datasets from ChinaFLUX sensor networks data stream.

  1. Null stream analysis of Pulsar Timing Array data: localisation of resolvable gravitational wave sources

    Science.gov (United States)

    Goldstein, Janna; Veitch, John; Sesana, Alberto; Vecchio, Alberto

    2018-04-01

    Super-massive black hole binaries are expected to produce a gravitational wave (GW) signal in the nano-Hertz frequency band which may be detected by pulsar timing arrays (PTAs) in the coming years. The signal is composed of both stochastic and individually resolvable components. Here we develop a generic Bayesian method for the analysis of resolvable sources based on the construction of `null-streams' which cancel the part of the signal held in common for each pulsar (the Earth-term). For an array of N pulsars there are N - 2 independent null-streams that cancel the GW signal from a particular sky location. This method is applied to the localisation of quasi-circular binaries undergoing adiabatic inspiral. We carry out a systematic investigation of the scaling of the localisation accuracy with signal strength and number of pulsars in the PTA. Additionally, we find that source sky localisation with the International PTA data release one is vastly superior than what is achieved by its constituent regional PTAs.

  2. The relationship between area deprivation and contact with community intellectual disability psychiatry.

    Science.gov (United States)

    Nicholson, L; Hotchin, H

    2015-05-01

    People with intellectual disabilities (ID) have high rates of psychiatric illness and are known to live in more deprived areas than the general population. This study investigated the relationship between area deprivation and contact with ID psychiatry. Psychiatric case notes and electronic records were used to identify all patients who had face-to-face contact with community ID psychiatric services over 1 year in the North East Community Health Partnership of Greater Glasgow and Clyde (estimated population 177,867). The Scottish Index of Multiple Deprivation (SIMD) were determined for the patient sample and for the general population living in the same area. Between 1 June 2012 and 1 June 2013, 184 patients were seen by ID psychiatry over a total of 553 contacts, with valid SIMD data for 179 patients and 543 contacts. Fifty-two per cent of patients (n = 93) lived in the most deprived SIMD decile, and 90.5% (n = 152) in the lowest 5 deciles. Compared with the general population, there were significantly more patients than expected living in the most deprived decile (Fisher's Exact test, P = 0.009) and in the most deprived 5 deciles (Fisher's Exact test, P = 0.001). The median number of contacts was 2 (interquartile range = 1-3). There was no significant association between the number of contacts and SIMD decile. Forty-eight point one per cent (n = 261) of all contacts were with patients living in the most deprived decile and 88.6% (n = 481) in the most deprived 5 deciles. This was significantly more than expected compared with general population data (Fisher's Exact test, P = 0.008 and Fisher's Exact test, P ≤ 0.001). In the area under study, contact with ID psychiatry was greater in more deprived areas. Given the high psychiatric morbidity of people with ID, if services do not adjust for deprivation, this may lead to further discrimination in an already disadvantaged population. © 2014 MENCAP and International Association of the Scientific Study of Intellectual

  3. The Stream-Catchment (StreamCat) and Lake-Catchment ...

    Science.gov (United States)

    Background/Question/MethodsLake and stream conditions respond to both natural and human-related landscape features. Characterizing these features within contributing areas (i.e., delineated watersheds) of streams and lakes could improve our understanding of how biological conditions vary spatially and improve the use, management, and restoration of these aquatic resources. However, the specialized geospatial techniques required to define and characterize stream and lake watersheds has limited their widespread use in both scientific and management efforts at large spatial scales. We developed the StreamCat and LakeCat Datasets to model, predict, and map the probable biological conditions of streams and lakes across the conterminous US (CONUS). Both StreamCat and LakeCat contain watershed-level characterizations of several hundred natural (e.g., soils, geology, climate, and land cover) and anthropogenic (e.g., urbanization, agriculture, mining, and forest management) landscape features for ca. 2.6 million stream segments and 376,000 lakes across the CONUS, respectively. These datasets can be paired with field samples to provide independent variables for modeling and other analyses. We paired 1,380 stream and 1,073 lake samples from the USEPAs National Aquatic Resource Surveys with StreamCat and LakeCat and used random forest (RF) to model and then map an invertebrate condition index and chlorophyll a concentration, respectively. Results/ConclusionsThe invertebrate

  4. Data analysis and pattern recognition in multiple databases

    CERN Document Server

    Adhikari, Animesh; Pedrycz, Witold

    2014-01-01

    Pattern recognition in data is a well known classical problem that falls under the ambit of data analysis. As we need to handle different data, the nature of patterns, their recognition and the types of data analyses are bound to change. Since the number of data collection channels increases in the recent time and becomes more diversified, many real-world data mining tasks can easily acquire multiple databases from various sources. In these cases, data mining becomes more challenging for several essential reasons. We may encounter sensitive data originating from different sources - those cannot be amalgamated. Even if we are allowed to place different data together, we are certainly not able to analyse them when local identities of patterns are required to be retained. Thus, pattern recognition in multiple databases gives rise to a suite of new, challenging problems different from those encountered before. Association rule mining, global pattern discovery, and mining patterns of select items provide different...

  5. Stream Response to an Extreme Defoliation Event

    Science.gov (United States)

    Gold, A.; Loffredo, J.; Addy, K.; Bernhardt, E. S.; Berdanier, A. B.; Schroth, A. W.; Inamdar, S. P.; Bowden, W. B.

    2017-12-01

    Extreme climatic events are known to profoundly impact stream flow and stream fluxes. These events can also exert controls on insect outbreaks, which may create marked changes in stream characteristics. The invasive Gypsy Moth (Lymantria dispar dispar) experiences episodic infestations based on extreme climatic conditions within the northeastern U.S. In most years, gypsy moth populations are kept in check by diseases. In 2016 - after successive years of unusually warm, dry spring and summer weather -gypsy moth caterpillars defoliated over half of Rhode Island's 160,000 forested ha. No defoliation of this magnitude had occurred for more than 30 years. We examined one RI headwater stream's response to the defoliation event in 2016 compared with comparable data in 2014 and 2015. Stream temperature and flow was gauged continuously by USGS and dissolved oxygen (DO) was measured with a YSI EXO2 sonde every 30 minutes during a series of deployments in the spring, summer and fall from 2014-2016. We used the single station, open channel method to estimate stream metabolism metrics. We also assessed local climate and stream temperature data from 2009-2016. We observed changes in stream responses during the defoliation event that suggest changes in ET, solar radiation and heat flux. Although the summer of 2016 had more drought stress (PDSI) than previous years, stream flow occurred throughout the summer, in contrast to several years with lower drought stress when stream flow ceased. Air temperature in 2016 was similar to prior years, but stream temperature was substantially higher than the prior seven years, likely due to the loss of canopy shading. DO declined dramatically in 2016 compared to prior years - more than the rising stream temperatures would indicate. Gross Primary Productivity was significantly higher during the year of the defoliation, indicating more total fixation of inorganic carbon from photo-autotrophs. In 2016, Ecosystem Respiration was also higher and Net

  6. Implicit Unstructured Aerodynamics on Emerging Multi- and Many-Core HPC Architectures

    KAUST Repository

    Al Farhan, Mohammed A.

    2017-03-13

    Shared memory parallelization of PETSc-FUN3D, an unstructured tetrahedral mesh Euler code previously characterized for distributed memory Single Program, Multiple Data (SPMD) for thousands of nodes, is hybridized with shared memory Single Instruction, Multiple Data (SIMD) for hundreds of threads per node. We explore thread-level performance optimizations on state-of-the-art multi- and many-core Intel processors, including the second generation of Xeon Phi, Knights Landing (KNL). We study the performance on the KNL with different configurations of memory and cluster modes, with code optimizations to minimize indirect addressing and enhance the cache locality. The optimizations employed are expected to be of value other unstructured applications as many-core architecture evolves.

  7. Real-time dissemination of air quality information using data streams and Web technologies: linking air quality to health risks in urban areas.

    Science.gov (United States)

    Davila, Silvije; Ilić, Jadranka Pečar; Bešlić, Ivan

    2015-06-01

    This article presents a new, original application of modern information and communication technology to provide effective real-time dissemination of air quality information and related health risks to the general public. Our on-line subsystem for urban real-time air quality monitoring is a crucial component of a more comprehensive integrated information system, which has been developed by the Institute for Medical Research and Occupational Health. It relies on a StreamInsight data stream management system and service-oriented architecture to process data streamed from seven monitoring stations across Zagreb. Parameters that are monitored include gases (NO, NO2, CO, O3, H2S, SO2, benzene, NH3), particulate matter (PM10 and PM2.5), and meteorological data (wind speed and direction, temperature and pressure). Streamed data are processed in real-time using complex continuous queries. They first go through automated validation, then hourly air quality index is calculated for every station, and a report sent to the Croatian Environment Agency. If the parameter values exceed the corresponding regulation limits for three consecutive hours, the web service generates an alert for population groups at risk. Coupled with the Common Air Quality Index model, our web application brings air pollution information closer to the general population and raises awareness about environmental and health issues. Soon we intend to expand the service to a mobile application that is being developed.

  8. Accelerated Adaptive MGS Phase Retrieval

    Science.gov (United States)

    Lam, Raymond K.; Ohara, Catherine M.; Green, Joseph J.; Bikkannavar, Siddarayappa A.; Basinger, Scott A.; Redding, David C.; Shi, Fang

    2011-01-01

    The Modified Gerchberg-Saxton (MGS) algorithm is an image-based wavefront-sensing method that can turn any science instrument focal plane into a wavefront sensor. MGS characterizes optical systems by estimating the wavefront errors in the exit pupil using only intensity images of a star or other point source of light. This innovative implementation of MGS significantly accelerates the MGS phase retrieval algorithm by using stream-processing hardware on conventional graphics cards. Stream processing is a relatively new, yet powerful, paradigm to allow parallel processing of certain applications that apply single instructions to multiple data (SIMD). These stream processors are designed specifically to support large-scale parallel computing on a single graphics chip. Computationally intensive algorithms, such as the Fast Fourier Transform (FFT), are particularly well suited for this computing environment. This high-speed version of MGS exploits commercially available hardware to accomplish the same objective in a fraction of the original time. The exploit involves performing matrix calculations in nVidia graphic cards. The graphical processor unit (GPU) is hardware that is specialized for computationally intensive, highly parallel computation. From the software perspective, a parallel programming model is used, called CUDA, to transparently scale multicore parallelism in hardware. This technology gives computationally intensive applications access to the processing power of the nVidia GPUs through a C/C++ programming interface. The AAMGS (Accelerated Adaptive MGS) software takes advantage of these advanced technologies, to accelerate the optical phase error characterization. With a single PC that contains four nVidia GTX-280 graphic cards, the new implementation can process four images simultaneously to produce a JWST (James Webb Space Telescope) wavefront measurement 60 times faster than the previous code.

  9. The effectiveness of streaming video on medical student learning: a case study.

    Science.gov (United States)

    Bridge, Patrick D; Jackson, Matt; Robinson, Leah

    2009-08-19

    Information technology helps meet today's medical students' needs by providing multiple curriculum delivery methods. Video streaming is an e-learning technology that uses the Internet to deliver curriculum while giving the student control of the content's delivery. There have been few studies conducted on the effectiveness of streaming video in medical schools. A 5-year retrospective study was conducted using three groups of students (n = 1736) to determine if the availability of streaming video in Years 1-2 of the basic science curriculum affected overall Step 1 scores for first-time test-takers. The results demonstrated a positive effect on program outcomes as streaming video became more readily available to students. Based on these findings, streaming video technology seems to be a viable tool to complement in-class delivery methods, to accommodate the needs of medical students, and to provide options for meeting the challenges of delivering the undergraduate medical curriculum. Further studies need to be conducted to continue validating the effectiveness of streaming video technology.

  10. Savannah River Laboratory hydrogeochemical and stream sediment reconnaissance. Preliminary raw data release, Charlotte 10 x 20 NTMS area, North Carolina and South Carolina. National Uranium Resource Evaluation Program

    International Nuclear Information System (INIS)

    Heffner, J.D.; Ferguson, R.B.

    1978-01-01

    This report presents preliminary results of stream sediment and ground water reconnaissance in the Charlotte National Topographic Map Series (NTMS) 1 0 x 2 0 quadrangle. Stream sediment samples were collected from small streams at 1254 sites for a nominal density of one site per 13 square kilometers (five square miles) in rural areas. Ground water samples were collected at 759 sites for a nominal density of one site per 25 square kilometers (ten squre miles). Neutron activation analysis (NAA) results are given for uranium and 16 other elements in sediments, and for uranium and 9 other elements in ground water. Field measurements and observations are reported for each site. Analytical data and field measurements are presented in tables and maps. Statistical summaries of data and a brief description of results are given. A generalized geologic map and a summary of the geology of the area are included. Key data are presented in page-sized hard copy. Supplementary data are on microfiche. Key data from stream sites include (1) water quality measurements (pH, conductivity, and alkalinity), (2) elements that may be related to potential uranium and thorium mineralization in this area (U, Th, Hf, Ce, and Dy), and (3) elements useful for geologic classification of the sample area (Ti, V, Fe, Mn, Al, and Sc). Supplementary data from stream sites include sample site descriptors (stream characteristics, vegetation, stream width, etc.) and additional elemental analyses that may be useful (F, Eu, Sm, La, Yb, and Lu). Key data from ground water sites include (1) water chemistry measurements (pH, conductivity, and alkalinity) and (2) elemental analyses (U, Na, Cl, Mg, Al, Mn, Br, V, and F). Supplementary data include site descriptors, information about the collection of the samples (well age, well depth, frequency of use of well, etc.), and analytical data for dysprosium

  11. Savannah River Laboratory Hydrogeochemical and Stream Sediment Reconnaissance. Preliminary raw data release: Spartanburg 10 x 20 NTMS area, North Carolina and South Carolina. National Uranium Resource Evaluation Program

    International Nuclear Information System (INIS)

    Heffner, J.D.; Ferguson, R.B.

    1977-12-01

    Preliminary results are presented of stream sediment and ground water reconnaissance in the Spartanburg National Topographic Map Series (NTMS) 1 0 x 2 0 quadrangle. Stream sediment samples were collected from small streams at 1202 sites for a nominal density of one site per 13 square kilometers (five square miles) in rural areas. Ground water samples were collected at 771 sites for a nominal density of one site per 25 square kilometers (ten square miles). Neutron activation analysis (NAA) results are given for uranium and 16 other elements in sediments, and for uranium and 9 other elements in ground water. Field measurements and observations are reported for each site. Analytical data and field measurements are presented in tables and maps. Statistical summaries of data and a brief description of results are given. A generalized geologic map and a summary of the geology of the area are included. Key data are presented in page-sized hard copy. Supplementary data are on microfiche. Key data from stream sites include (1) water quality measurements (pH, conductivity, and alkalinity), (2) elements that may be related to potential uranium and thorium mineralization in this area (U, Th, Hf, Ce, and Dy), and (3) elements useful for geologic classification of the sample area (Ti, V, Fe, Mn, A, and Sc). Supplementary data from stream sites include sample site descriptors (stream characteristics, vegetation, stream width, etc.) and additional elemental analyses that may be useful (F, Eu, Sm, La, Yb, and Lu). Key data from ground water sites include (1) water chemistry measurements (pH, conductivity, and alkalinity) and (2) elemental analyses (U, Na, Cl, Mg, Al, Mn, Br, V, and F). Supplementary data include site descriptors, information about the collection of the samples (well age, well depth, frequency of use of well, etc.), and analytical data for dyprosium

  12. Performing Frame Transformations to Correctly Stream Position Data

    Science.gov (United States)

    Franco, Tom

    Unmanned Aerial Vehicles (UAV) are starting to become a more common occurrence today. What started off as highly classified military weapons with little known information, have become part of everyday life for the common individual. UAV's still carry a great deal of importance in paving the way for unmanned flight. UAV's hold major potential for many amazing technological advances within the near future. Drones have become such a common backyard toy for individuals all over the world as well as the way of the future. Major corporations, such as Amazon, are starting to test drones for delivering small packages. Uber has stated that they want to get to the point where cars will be self-driving, already implementing their testing facility for self-driving cars. It is crazy to think that if an order from amazon is processed, it could arrive at the desired destination the same day within minutes of being processed. To get to that point, there is a lot to consider. First, and most importantly, the drone must be largely autonomous with no minimal human control. The drone also must be able to communicate effectively and relay its position to some sort of tracking device, whether it be a GPS signal or software. How would it go about this? What sort of factors make this possible fantasy of the future a tangible reality? The drone must communicate with numerous devices, be in the proper orientation and have the data being streamed be associated with the proper direction. Since there are a variety of potential directions for the drone to move, odds are there will be some sort of data conversion involved. When testing turbomachinery, sensors used to be placed on the rotating piece of machinery and frame transformations were done to relay the data from the rotating frame to that of the inertial frame. Using this concept, exploring the use of frame transformations to relay position data is conducted. Once explore, testing can be conducted to collect data and once the data is

  13. Interaction between stream temperature, streamflow, and groundwater exchanges in alpine streams

    Science.gov (United States)

    Constantz, James E.

    1998-01-01

    Four alpine streams were monitored to continuously collect stream temperature and streamflow for periods ranging from a week to a year. In a small stream in the Colorado Rockies, diurnal variations in both stream temperature and streamflow were significantly greater in losing reaches than in gaining reaches, with minimum streamflow losses occurring early in the day and maximum losses occurring early in the evening. Using measured stream temperature changes, diurnal streambed infiltration rates were predicted to increase as much as 35% during the day (based on a heat and water transport groundwater model), while the measured increase in streamflow loss was 40%. For two large streams in the Sierra Nevada Mountains, annual stream temperature variations ranged from 0° to 25°C. In summer months, diurnal stream temperature variations were 30–40% of annual stream temperature variations, owing to reduced streamflows and increased atmospheric heating. Previous reports document that one Sierra stream site generally gains groundwater during low flows, while the second Sierra stream site may lose water during low flows. For August the diurnal streamflow variation was 11% at the gaining stream site and 30% at the losing stream site. On the basis of measured diurnal stream temperature variations, streambed infiltration rates were predicted to vary diurnally as much as 20% at the losing stream site. Analysis of results suggests that evapotranspiration losses determined diurnal streamflow variations in the gaining reaches, while in the losing reaches, evapotranspiration losses were compounded by diurnal variations in streambed infiltration. Diurnal variations in stream temperature were reduced in the gaining reaches as a result of discharging groundwater of relatively constant temperature. For the Sierra sites, comparison of results with those from a small tributary demonstrated that stream temperature patterns were useful in delineating discharges of bank storage following

  14. Disintegration of a marine-based ice stream - evidence from the Norwegian Channel, north-eastern North Sea

    Science.gov (United States)

    Morén, Björn M.; Petter Sejrup, Hans; Hjelstuen, Berit O.; Haflidason, Haflidi; Schäuble, Cathrina; Borge, Marianne

    2014-05-01

    The Norwegian Channel Ice Stream repeatedly drained large part of the Fennoscandian Ice Sheet through Mid and Late Pleistocene glacial stages. During parts of Marine Isotope Stages 2 and 3, glacial ice from Fennoscandia and the British Isles coalesced in the central North Sea and the Norwegian Channel Ice Stream reached the shelf edge on multiple occasions. Through the last decades a large amount of acoustic and sediment core data have been collected from the Norwegian Channel, providing a good background for studies focussing on stability- and development-controlling parameters for marine-based ice streams, the retreat rate of the Norwegian Channel Ice Stream, and the behaviour of the Fennoscandian Ice Sheet. Further, this improved understanding can be used to develop more accurate numerical climate models and models which can be used to model ice-sheet behaviour of the past as well as the future. This study presents new acoustic records and data from sediment cores which contribute to a better understanding of the retreat pattern and the retreat rate of the last ice stream that occupied the Norwegian Channel. From bathymetric and TOPAS seismic data, mega-scale glacial lineations, grounding-zone wedges, and end moraines have been mapped, thereby allowing us to reconstruct the pro- and subglacial conditions at the time of the creation of these landforms. It is concluded that the whole Norwegian Channel was deglaciated in just over 1 000 years and that for most of this time the ice margin was located at positions reflected by depositional grounding-zone wedges. Further work will explore the influence of channel shape and feeding of ice from western Norwegian fjords on this retreat pattern through numerical modelling.

  15. Network Monitoring as a Streaming Analytics Problem

    KAUST Repository

    Gupta, Arpit

    2016-11-02

    Programmable switches make it easier to perform flexible network monitoring queries at line rate, and scalable stream processors make it possible to fuse data streams to answer more sophisticated queries about the network in real-time. Unfortunately, processing such network monitoring queries at high traffic rates requires both the switches and the stream processors to filter the traffic iteratively and adaptively so as to extract only that traffic that is of interest to the query at hand. Others have network monitoring in the context of streaming; yet, previous work has not closed the loop in a way that allows network operators to perform streaming analytics for network monitoring applications at scale. To achieve this objective, Sonata allows operators to express a network monitoring query by considering each packet as a tuple and efficiently partitioning each query between the switches and the stream processor through iterative refinement. Sonata extracts only the traffic that pertains to each query, ensuring that the stream processor can scale traffic rates of several terabits per second. We show with a simple example query involving DNS reflection attacks and traffic traces from one of the world\\'s largest IXPs that Sonata can capture 95% of all traffic pertaining to the query, while reducing the overall data rate by a factor of about 400 and the number of required counters by four orders of magnitude. Copyright 2016 ACM.

  16. Various aspects of vehicles image data-streams reduction for road traffic sufficient description

    Directory of Open Access Journals (Sweden)

    Jan PIECHA

    2007-01-01

    Full Text Available The on-line image processing was implemented for video-camera usage for traffic control. Due to reduce the immense data sets dimension various speculations of data sampling methods were introduced. At the beginning the needed sampling ratio has been found then simple but effective image processing algorithms have to be chosen, finally the hardware solutions for parallel processing are discussed. The PLA computing engine was involved for coping with this task; for fulfilling the assumed characteristics. The developer has to consider several restrictions and preferences. None universal algorithm is available up to now. The reported works, concern vehicles stream recorders development that has to do all recording and computing procedures in strictly defined time limits.

  17. Identify the dominant variables to predict stream water temperature

    Science.gov (United States)

    Chien, H.; Flagler, J.

    2016-12-01

    Stream water temperature is a critical variable controlling water quality and the health of aquatic ecosystems. Accurate prediction of water temperature and the assessment of the impacts of environmental variables on water temperature variation are critical for water resources management, particularly in the context of water quality and aquatic ecosystem sustainability. The objective of this study is to measure stream water temperature and air temperature and to examine the importance of streamflow on stream water temperature prediction. The measured stream water temperature and air temperature will be used to test two hypotheses: 1) streamflow is a relatively more important factor than air temperature in regulating water temperature, and 2) by combining air temperature and streamflow data stream water temperature can be more accurately estimated. Water and air temperature data loggers are placed at two USGS stream gauge stations #01362357and #01362370, located in the upper Esopus Creek watershed in Phonecia, NY. The ARIMA (autoregressive integrated moving average) time series model is used to analyze the measured water temperature data, identify the dominant environmental variables, and predict the water temperature with identified dominant variable. The preliminary results show that streamflow is not a significant variable in predicting stream water temperature at both USGS gauge stations. Daily mean air temperature is sufficient to predict stream water temperature at this site scale.

  18. Online Censoring for Large-Scale Regressions with Application to Streaming Big Data.

    Science.gov (United States)

    Berberidis, Dimitris; Kekatos, Vassilis; Giannakis, Georgios B

    2016-08-01

    On par with data-intensive applications, the sheer size of modern linear regression problems creates an ever-growing demand for efficient solvers. Fortunately, a significant percentage of the data accrued can be omitted while maintaining a certain quality of statistical inference with an affordable computational budget. This work introduces means of identifying and omitting less informative observations in an online and data-adaptive fashion. Given streaming data, the related maximum-likelihood estimator is sequentially found using first- and second-order stochastic approximation algorithms. These schemes are well suited when data are inherently censored or when the aim is to save communication overhead in decentralized learning setups. In a different operational scenario, the task of joint censoring and estimation is put forth to solve large-scale linear regressions in a centralized setup. Novel online algorithms are developed enjoying simple closed-form updates and provable (non)asymptotic convergence guarantees. To attain desired censoring patterns and levels of dimensionality reduction, thresholding rules are investigated too. Numerical tests on real and synthetic datasets corroborate the efficacy of the proposed data-adaptive methods compared to data-agnostic random projection-based alternatives.

  19. Learning From Short Text Streams With Topic Drifts.

    Science.gov (United States)

    Li, Peipei; He, Lu; Wang, Haiyan; Hu, Xuegang; Zhang, Yuhong; Li, Lei; Wu, Xindong

    2017-09-18

    Short text streams such as search snippets and micro blogs have been popular on the Web with the emergence of social media. Unlike traditional normal text streams, these data present the characteristics of short length, weak signal, high volume, high velocity, topic drift, etc. Short text stream classification is hence a very challenging and significant task. However, this challenge has received little attention from the research community. Therefore, a new feature extension approach is proposed for short text stream classification with the help of a large-scale semantic network obtained from a Web corpus. It is built on an incremental ensemble classification model for efficiency. First, more semantic contexts based on the senses of terms in short texts are introduced to make up of the data sparsity using the open semantic network, in which all terms are disambiguated by their semantics to reduce the noise impact. Second, a concept cluster-based topic drifting detection method is proposed to effectively track hidden topic drifts. Finally, extensive studies demonstrate that as compared to several well-known concept drifting detection methods in data stream, our approach can detect topic drifts effectively, and it enables handling short text streams effectively while maintaining the efficiency as compared to several state-of-the-art short text classification approaches.

  20. System for processing an encrypted instruction stream in hardware

    Science.gov (United States)

    Griswold, Richard L.; Nickless, William K.; Conrad, Ryan C.

    2016-04-12

    A system and method of processing an encrypted instruction stream in hardware is disclosed. Main memory stores the encrypted instruction stream and unencrypted data. A central processing unit (CPU) is operatively coupled to the main memory. A decryptor is operatively coupled to the main memory and located within the CPU. The decryptor decrypts the encrypted instruction stream upon receipt of an instruction fetch signal from a CPU core. Unencrypted data is passed through to the CPU core without decryption upon receipt of a data fetch signal.

  1. Assimilation of global versus local data sets into a regional model of the Gulf Stream system. 1. Data effectiveness

    Science.gov (United States)

    Malanotte-Rizzoli, Paola; Young, Roberta E.

    1995-12-01

    The primary objective of this paper is to assess the relative effectiveness of data sets with different space coverage and time resolution when they are assimilated into an ocean circulation model. We focus on obtaining realistic numerical simulations of the Gulf Stream system typically of the order of 3-month duration by constructing a "synthetic" ocean simultaneously consistent with the model dynamics and the observations. The model used is the Semispectral Primitive Equation Model. The data sets are the "global" Optimal Thermal Interpolation Scheme (OTIS) 3 of the Fleet Numerical Oceanography Center providing temperature and salinity fields with global coverage and with bi-weekly frequency, and the localized measurements, mostly of current velocities, from the central and eastern array moorings of the Synoptic Ocean Prediction (SYNOP) program, with daily frequency but with a very small spatial coverage. We use a suboptimal assimilation technique ("nudging"). Even though this technique has already been used in idealized data assimilation studies, to our knowledge this is the first study in which the effectiveness of nudging is tested by assimilating real observations of the interior temperature and salinity fields. This is also the first work in which a systematic assimilation is carried out of the localized, high-quality SYNOP data sets in numerical experiments longer than 1-2 weeks, that is, not aimed to forecasting. We assimilate (1) the global OTIS 3 alone, (2) the local SYNOP observations alone, and (3) both OTIS 3 and SYNOP observations. We assess the success of the assimilations with quantitative measures of performance, both on the global and local scale. The results can be summarized as follows. The intermittent assimilation of the global OTIS 3 is necessary to keep the model "on track" over 3-month simulations on the global scale. As OTIS 3 is assimilated at every model grid point, a "gentle" weight must be prescribed to it so as not to overconstrain

  2. Iowa Flood Information System: Towards Integrated Data Management, Analysis and Visualization

    Science.gov (United States)

    Demir, I.; Krajewski, W. F.; Goska, R.; Mantilla, R.; Weber, L. J.; Young, N.

    2012-04-01

    The Iowa Flood Information System (IFIS) is a web-based platform developed by the Iowa Flood Center (IFC) to provide access to flood inundation maps, real-time flood conditions, flood forecasts both short-term and seasonal, flood-related data, information and interactive visualizations for communities in Iowa. The key element of the system's architecture is the notion of community. Locations of the communities, those near streams and rivers, define basin boundaries. The IFIS provides community-centric watershed and river characteristics, weather (rainfall) conditions, and streamflow data and visualization tools. Interactive interfaces allow access to inundation maps for different stage and return period values, and flooding scenarios with contributions from multiple rivers. Real-time and historical data of water levels, gauge heights, and rainfall conditions are available in the IFIS by streaming data from automated IFC bridge sensors, USGS stream gauges, NEXRAD radars, and NWS forecasts. Simple 2D and 3D interactive visualizations in the IFIS make the data more understandable to general public. Users are able to filter data sources for their communities and selected rivers. The data and information on IFIS is also accessible through web services and mobile applications. The IFIS is optimized for various browsers and screen sizes to provide access through multiple platforms including tablets and mobile devices. The IFIS includes a rainfall-runoff forecast model to provide a five-day flood risk estimate for around 500 communities in Iowa. Multiple view modes in the IFIS accommodate different user types from general public to researchers and decision makers by providing different level of tools and details. River view mode allows users to visualize data from multiple IFC bridge sensors and USGS stream gauges to follow flooding condition along a river. The IFIS will help communities make better-informed decisions on the occurrence of floods, and will alert communities

  3. Mining gene expression data of multiple sclerosis.

    Directory of Open Access Journals (Sweden)

    Pi Guo

    Full Text Available Microarray produces a large amount of gene expression data, containing various biological implications. The challenge is to detect a panel of discriminative genes associated with disease. This study proposed a robust classification model for gene selection using gene expression data, and performed an analysis to identify disease-related genes using multiple sclerosis as an example.Gene expression profiles based on the transcriptome of peripheral blood mononuclear cells from a total of 44 samples from 26 multiple sclerosis patients and 18 individuals with other neurological diseases (control were analyzed. Feature selection algorithms including Support Vector Machine based on Recursive Feature Elimination, Receiver Operating Characteristic Curve, and Boruta algorithms were jointly performed to select candidate genes associating with multiple sclerosis. Multiple classification models categorized samples into two different groups based on the identified genes. Models' performance was evaluated using cross-validation methods, and an optimal classifier for gene selection was determined.An overlapping feature set was identified consisting of 8 genes that were differentially expressed between the two phenotype groups. The genes were significantly associated with the pathways of apoptosis and cytokine-cytokine receptor interaction. TNFSF10 was significantly associated with multiple sclerosis. A Support Vector Machine model was established based on the featured genes and gave a practical accuracy of ∼86%. This binary classification model also outperformed the other models in terms of Sensitivity, Specificity and F1 score.The combined analytical framework integrating feature ranking algorithms and Support Vector Machine model could be used for selecting genes for other diseases.

  4. ATLAS Live: Collaborative Information Streams

    CERN Document Server

    Goldfarb, S; The ATLAS collaboration

    2010-01-01

    I report on a pilot project launched in 2010 focusing on facilitating communication and information exchange within the ATLAS Collaboration, through the combination of digital signage software and webcasting. The project, called ATLAS Live, implements video streams of information, ranging from detailed detector and data status to educational and outreach material. The content, including text, images, video and audio, is collected, visualised and scheduled using the SCALA digital signage software system. The system is robust and flexible, allowing for the usage of scripts to input data from remote sources, such as the CERN Document Server, Indico, or any available URL, and to integrate these sources into professional-quality streams, including text scrolling, transition effects, inter and intrascreen divisibility. The video is made available to the collaboration or public through the encoding and webcasting of standard video streams, viewable on all common platforms, using a web browser or other common video t...

  5. Relation between Streaming Potential and Streaming Electrification Generated by Streaming of Water through a Sandwich-type Cell

    OpenAIRE

    Maruyama, Kazunori; Nikaido, Mitsuru; Hara, Yoshinori; Tanizaki, Yoshie

    2012-01-01

    Both streaming potential and accumulated charge of water flowed out were measured simultaneously using a sandwich-type cell. The voltages generated in divided sections along flow direction satisfied additivity. The sign of streaming potential agreed with that of streaming electrification. The relation between streaming potential and streaming electrification was explained from a viewpoint of electrical double layer in glass-water interface.

  6. Synchronized analysis of testbeam data with the Judith software

    CERN Document Server

    McGoldrick, Garrin; Gorišek, Andrej

    2014-01-01

    The Judith software performs pixel detector analysis tasks utilizing two different data streams such as those produced by the reference and tested devices typically found in a testbeam. This software addresses and fixes problems arising from the desynchronization of the two simultaneously triggered data streams by detecting missed triggers in either of the streams. The software can perform all tasks required to generate particle tracks using multiple detector planes: it can align the planes, cluster hits and generate tracks from these clusters. This information can then be used to measure the properties of a particle detector with very fine spatial resolution. It was tested at DESY in the Kartel telescope, a silicon tracking detector, with ATLAS Diamond Beam Monitor modules as a device under test.

  7. StreamStats in Oklahoma - Drainage-Basin Characteristics and Peak-Flow Frequency Statistics for Ungaged Streams

    Science.gov (United States)

    Smith, S. Jerrod; Esralew, Rachel A.

    2010-01-01

    The USGS Streamflow Statistics (StreamStats) Program was created to make geographic information systems-based estimation of streamflow statistics easier, faster, and more consistent than previously used manual techniques. The StreamStats user interface is a map-based internet application that allows users to easily obtain streamflow statistics, basin characteristics, and other information for user-selected U.S. Geological Survey data-collection stations and ungaged sites of interest. The application relies on the data collected at U.S. Geological Survey streamflow-gaging stations, computer aided computations of drainage-basin characteristics, and published regression equations for several geographic regions comprising the United States. The StreamStats application interface allows the user to (1) obtain information on features in selected map layers, (2) delineate drainage basins for ungaged sites, (3) download drainage-basin polygons to a shapefile, (4) compute selected basin characteristics for delineated drainage basins, (5) estimate selected streamflow statistics for ungaged points on a stream, (6) print map views, (7) retrieve information for U.S. Geological Survey streamflow-gaging stations, and (8) get help on using StreamStats. StreamStats was designed for national application, with each state, territory, or group of states responsible for creating unique geospatial datasets and regression equations to compute selected streamflow statistics. With the cooperation of the Oklahoma Department of Transportation, StreamStats has been implemented for Oklahoma and is available at http://water.usgs.gov/osw/streamstats/. The Oklahoma StreamStats application covers 69 processed hydrologic units and most of the state of Oklahoma. Basin characteristics available for computation include contributing drainage area, contributing drainage area that is unregulated by Natural Resources Conservation Service floodwater retarding structures, mean-annual precipitation at the

  8. Efficient Delivery of Scalable Video Using a Streaming Class Model

    Directory of Open Access Journals (Sweden)

    Jason J. Quinlan

    2018-03-01

    Full Text Available When we couple the rise in video streaming with the growing number of portable devices (smart phones, tablets, laptops, we see an ever-increasing demand for high-definition video online while on the move. Wireless networks are inherently characterised by restricted shared bandwidth and relatively high error loss rates, thus presenting a challenge for the efficient delivery of high quality video. Additionally, mobile devices can support/demand a range of video resolutions and qualities. This demand for mobile streaming highlights the need for adaptive video streaming schemes that can adjust to available bandwidth and heterogeneity, and can provide a graceful changes in video quality, all while respecting viewing satisfaction. In this context, the use of well-known scalable/layered media streaming techniques, commonly known as scalable video coding (SVC, is an attractive solution. SVC encodes a number of video quality levels within a single media stream. This has been shown to be an especially effective and efficient solution, but it fares badly in the presence of datagram losses. While multiple description coding (MDC can reduce the effects of packet loss on scalable video delivery, the increased delivery cost is counterproductive for constrained networks. This situation is accentuated in cases where only the lower quality level is required. In this paper, we assess these issues and propose a new approach called Streaming Classes (SC through which we can define a key set of quality levels, each of which can be delivered in a self-contained manner. This facilitates efficient delivery, yielding reduced transmission byte-cost for devices requiring lower quality, relative to MDC and Adaptive Layer Distribution (ALD (42% and 76% respective reduction for layer 2, while also maintaining high levels of consistent quality. We also illustrate how selective packetisation technique can further reduce the effects of packet loss on viewable quality by

  9. Experimental Analysis of Cell Function Using Cytoplasmic Streaming

    Science.gov (United States)

    Janssens, Peter; Waldhuber, Megan

    2012-01-01

    This laboratory exercise investigates the phenomenon of cytoplasmic streaming in the fresh water alga "Nitella". Students use the fungal toxin cytochalasin D, an inhibitor of actin polymerization, to investigate the mechanism of streaming. Students use simple statistical methods to analyze their data. Typical student data are provided. (Contains 3…

  10. Seasonal variability of stream water quality response to storm events captured using high-frequency and multi-parameter data

    Science.gov (United States)

    Fovet, O.; Humbert, G.; Dupas, R.; Gascuel-Odoux, C.; Gruau, G.; Jaffrezic, A.; Thelusma, G.; Faucheux, M.; Gilliet, N.; Hamon, Y.; Grimaldi, C.

    2018-04-01

    The response of stream chemistry to storm is of major interest for understanding the export of dissolved and particulate species from catchments. The related challenge is the identification of active hydrological flow paths during these events and of the sources of chemical elements for which these events are hot moments of exports. An original four-year data set that combines high frequency records of stream flow, turbidity, nitrate and dissolved organic carbon concentrations, and piezometric levels was used to characterize storm responses in a headwater agricultural catchment. The data set was used to test to which extend the shallow groundwater was impacting the variability of storm responses. A total of 177 events were described using a set of quantitative and functional descriptors related to precipitation, stream and groundwater pre-event status and event dynamics, and to the relative dynamics between water quality parameters and flow via hysteresis indices. This approach led to identify different types of response for each water quality parameter which occurrence can be quantified and related to the seasonal functioning of the catchment. This study demonstrates that high-frequency records of water quality are precious tools to study/unique in their ability to emphasize the variability of catchment storm responses.

  11. In-stream Nitrogen Processing and Dilution in an Agricultural Stream Network

    Science.gov (United States)

    Prior, K.; Ward, A. S.; Davis, C. A.; Burgin, A. J.; Loecke, T.; Riveros-Iregui, D. A.; Thomas, S. A.; St Clair, M. A.

    2014-12-01

    The interaction of agricultural fertilizer use and extremes in drought and flood conditions in 2012-2013 set up conditions for a natural experiment on watershed-scale nutrient dynamics. The region-wide drought in 2012 left surface soils disconnected from stream networks and restricted nutrient use by crops, resulting in an unusually large nitrogen pool in soil columns through the winter. When wet conditions returned to the Midwest in 2013, the unused fertilizer was mobilized, resulting in a six-week period of extremely high in-stream nutrient concentrations. This study analyses three synoptic samples from the Iowa-Cedar River Basin in 2013 to quantify patterns in nitrogen dynamics. We use multiple conservative ions as tracers to estimate dilution by lateral inflows. We also estimate nutrient spiraling metrics by treating the fertilizer pulse as a constant rate nutrient addition across the watershed—a scale on which these processes are increasingly modeled numerically, but on which standard nutrient addition experiments are simply not feasible. Results of this study compare patterns in dilution and uptake across spatial and temporal scales, and bound feasible explanations for each reach of the network.

  12. Data report: Illinois, Indiana, Kentucky, Tennessee, and Ohio. National Uranium Resource Evaluation Program. Hydrogeochemical and stream sediment reconnaissance

    International Nuclear Information System (INIS)

    Sargent, K.A.; Cook, J.R.; Fay, W.M.

    1982-02-01

    This report presents the results of ground water, stream water, and stream sediment reconnaissance in Illinois, Indiana, Kentucky, Tennessee, and Ohio. The following sample types were collected in each state: Illinois - 716 stream sediment, 1046 ground water, 337 stream water; Indiana - 126 stream sediment, 443 ground water, 111 stream water; Kentucky - 4901 stream sediment, 6408 ground water, 3966 stream water; Tennessee - 3309 stream sediment, 3574 ground water, 1584 stream water; Ohio - 1214 stream sediment, 2049 ground water, 1205 stream water. Neutron activation analyses are given for U, Br, Cl, F, Mn, Na, Al, V, and Dy in ground water and stream water, and for U, Th, Hf, Ce, Fe, Mn, Na, Sc, Ti, V, Al, Dy, Eu, La, Sm, Yb, and Lu in sediments. Supplementary analyses by other techniques are reported for U (extractable), Ag, As, Ba, Be, Ca, Co, Cr, Cu, K, Li, Mg, Mo, Nb, Ni, P, Pb, Se, Sn, Sr, W, Y, and Zn. These analyses were made on 248 sediment samples from Tennessee. Field measurements and observations are reported for each site. Oak Ridge National Laboratory analyzed sediment samples which were not analyzed by Savannah River Laboratory neutron activation

  13. The multiple imputation method: a case study involving secondary data analysis.

    Science.gov (United States)

    Walani, Salimah R; Cleland, Charles M

    2015-05-01

    To illustrate with the example of a secondary data analysis study the use of the multiple imputation method to replace missing data. Most large public datasets have missing data, which need to be handled by researchers conducting secondary data analysis studies. Multiple imputation is a technique widely used to replace missing values while preserving the sample size and sampling variability of the data. The 2004 National Sample Survey of Registered Nurses. The authors created a model to impute missing values using the chained equation method. They used imputation diagnostics procedures and conducted regression analysis of imputed data to determine the differences between the log hourly wages of internationally educated and US-educated registered nurses. The authors used multiple imputation procedures to replace missing values in a large dataset with 29,059 observations. Five multiple imputed datasets were created. Imputation diagnostics using time series and density plots showed that imputation was successful. The authors also present an example of the use of multiple imputed datasets to conduct regression analysis to answer a substantive research question. Multiple imputation is a powerful technique for imputing missing values in large datasets while preserving the sample size and variance of the data. Even though the chained equation method involves complex statistical computations, recent innovations in software and computation have made it possible for researchers to conduct this technique on large datasets. The authors recommend nurse researchers use multiple imputation methods for handling missing data to improve the statistical power and external validity of their studies.

  14. The influence of multiple chemical and non-chemical stressors on benthic communities in a mid-west agricultural stream.

    Science.gov (United States)

    Hall, Lenwood W; Killen, Willian D; Anderson, Ronald D; Alden, Raymond W

    2017-08-24

    The objective of this 3-year study was to characterize benthic communities and physical habitat in an agricultural stream in the mid-west area of the United States (Big Bureau Creek, Illinois). Concurrent basic water quality parameters and seven nutrients were measured in the water column. Sediment measurements from depositional areas were conducted for bifenthrin, Total Organic Carbon, grain size, polychlorinated biphenyls (PCBs) and eight metals. All parameters were measured at 12 sites annually during the late summer for a 3-year period (2014, 2015 and 2016). Univariate regressions, stepwise multiple regressions and canonical correlation statistical analyses were used to determine the relationship between various benthic metrics (i.e., taxa richness and abundance) and all the measured parameters for the 3-year database. Benthic communities comprising 108-110 taxa were collected annually, and were generally dominated by sensitive caddisflies and mayflies. These communities were rated as good to exceptional using the Ohio Invertebrate Community Index. Physical habitat for the various sites was rated as good using the Ohio Qualitative Habitat Evaluation Index, thus suggesting that habitat is not a significant stressor that would likely impact resident benthic communities. Based on a comparison of measured in-stream total nitrogen and total phosphorus concentrations and criterion value exceedances, it appears that the in-stream nutrient concentrations could be potentially stressful to resident benthic biota. Metal concentrations were below established NOAA Threshold Effects Levels at all sites. Measured PCB concentrations were below levels of detection at all sites. Toxic units' (TUs) calculations based on using sensitive laboratory strains of Hyalella were less than 0.1 for bifenthrin, thus suggesting that bifenthrin sediment toxicity was unlikely. Thirty significant relationships reported between benthic metrics and the various environmental variables based on the

  15. VPipe: Virtual Pipelining for Scheduling of DAG Stream Query Plans

    Science.gov (United States)

    Wang, Song; Gupta, Chetan; Mehta, Abhay

    There are data streams all around us that can be harnessed for tremendous business and personal advantage. For an enterprise-level stream processing system such as CHAOS [1] (Continuous, Heterogeneous Analytic Over Streams), handling of complex query plans with resource constraints is challenging. While several scheduling strategies exist for stream processing, efficient scheduling of complex DAG query plans is still largely unsolved. In this paper, we propose a novel execution scheme for scheduling complex directed acyclic graph (DAG) query plans with meta-data enriched stream tuples. Our solution, called Virtual Pipelined Chain (or VPipe Chain for short), effectively extends the "Chain" pipelining scheduling approach to complex DAG query plans.

  16. Multiple Scattering Principal Component-based Radiative Transfer Model (PCRTM) from Far IR to UV-Vis

    Science.gov (United States)

    Liu, X.; Wu, W.; Yang, Q.

    2017-12-01

    Modern satellite hyperspectral satellite remote sensors such as AIRS, CrIS, IASI, CLARREO all require accurate and fast radiative transfer models that can deal with multiple scattering of clouds and aerosols to explore the information contents. However, performing full radiative transfer calculations using multiple stream methods such as discrete ordinate (DISORT), doubling and adding (AD), successive order of scattering order of scattering (SOS) are very time consuming. We have developed a principal component-based radiative transfer model (PCRTM) to reduce the computational burden by orders of magnitudes while maintain high accuracy. By exploring spectral correlations, the PCRTM reduce the number of radiative transfer calculations in frequency domain. It further uses a hybrid stream method to decrease the number of calls to the computational expensive multiple scattering calculations with high stream numbers. Other fast parameterizations have been used in the infrared spectral region reduce the computational time to milliseconds for an AIRS forward simulation (2378 spectral channels). The PCRTM has been development to cover spectral range from far IR to UV-Vis. The PCRTM model have been be used for satellite data inversions, proxy data generation, inter-satellite calibrations, spectral fingerprinting, and climate OSSE. We will show examples of applying the PCRTM to single field of view cloudy retrievals of atmospheric temperature, moisture, traces gases, clouds, and surface parameters. We will also show how the PCRTM are used for the NASA CLARREO project.

  17. A stream cipher based on a spatiotemporal chaotic system

    International Nuclear Information System (INIS)

    Li Ping; Li Zhong; Halang, Wolfgang A.; Chen Guanrong

    2007-01-01

    A stream cipher based on a spatiotemporal chaotic system is proposed. A one-way coupled map lattice consisting of logistic maps is served as the spatiotemporal chaotic system. Multiple keystreams are generated from the coupled map lattice by using simple algebraic computations, and then are used to encrypt plaintext via bitwise XOR. These make the cipher rather simple and efficient. Numerical investigation shows that the cryptographic properties of the generated keystream are satisfactory. The cipher seems to have higher security, higher efficiency and lower computation expense than the stream cipher based on a spatiotemporal chaotic system proposed recently

  18. Cellular Subcompartments through Cytoplasmic Streaming.

    Science.gov (United States)

    Pieuchot, Laurent; Lai, Julian; Loh, Rachel Ann; Leong, Fong Yew; Chiam, Keng-Hwee; Stajich, Jason; Jedd, Gregory

    2015-08-24

    Cytoplasmic streaming occurs in diverse cell types, where it generally serves a transport function. Here, we examine streaming in multicellular fungal hyphae and identify an additional function wherein regimented streaming forms distinct cytoplasmic subcompartments. In the hypha, cytoplasm flows directionally from cell to cell through septal pores. Using live-cell imaging and computer simulations, we identify a flow pattern that produces vortices (eddies) on the upstream side of the septum. Nuclei can be immobilized in these microfluidic eddies, where they form multinucleate aggregates and accumulate foci of the HDA-2 histone deacetylase-associated factor, SPA-19. Pores experiencing flow degenerate in the absence of SPA-19, suggesting that eddy-trapped nuclei function to reinforce the septum. Together, our data show that eddies comprise a subcellular niche favoring nuclear differentiation and that subcompartments can be self-organized as a consequence of regimented cytoplasmic streaming. Copyright © 2015 Elsevier Inc. All rights reserved.

  19. The effects of logging road construction on insect drop into a small coastal stream

    Science.gov (United States)

    Lloyd J. Hess

    1969-01-01

    Abstract - Because stream fisheries are so closely associated with forested watersheds, it is necessary that the streams and forests be managed jointly under a system of multiple use. This requires a knowledge of the interrelationships between these resources to yield maximum returns from both. It is the purpose of this paper to relate logging practices to fish...

  20. Savannah River Laboratory hydrogeochemical and stream sediment reconnaissance. Preliminary raw data release: Greenville 10 x 20 NTMS area Georgia, North Carolina, and South Carolina. National Uranium Resource Evaluation Program

    International Nuclear Information System (INIS)

    Ferguson, R.B.

    1978-03-01

    Preliminary results of stream sediment and ground water reconnaissance in the Greenville National Topographic Map Series (NTMS) 1 0 x 2 0 quadrangle are presented. Stream sediment samples were collected from small streams at 1413 sites for a nominal density of one site per 13 square kilometers in rural areas. Ground water samples were collected at 731 sites for a nominal density of one site per 25 square kilometers. Neutron activation analysis (NAA) results are given for uranium and 16 other elements in sediments, and for uranium and 9 other elements in ground water. Field measurements and observations are reported for each site. Analytical data and field measurements are presented in tables and maps. Statistical summaries of data and a brief description of results are given. A generalized geologic map and a summary of the geology of the area are included. Key data are presented in page-sized hard copy. Supplementary data are on microfiche. Key data from stream sites include (1) water quality measurements (pH, conductivity, and alkalinity), (2) elements that may be related to potential uranium and thorium mineralization in this area (U, Th, Hf, Ce, and Dy) and (3) elements useful for geologic classification of the sample area (Ti, V, Fe, Mn, Al, and Sc). Supplementary data from stream sites include sample site descriptors (stream characteristics, vegetation, stream width, etc.) and additional elemental analyses that may be useful (F, Eu, Sm, La, Yb, and Lu). Key data from ground water sites include (1) water chemistry measurements (pH, conductivity, and alkalinity) and (2) elemental analyses (U, Na, Cl, Mg, Al, Mn, Br, V, and F). Supplementary data include site descriptors, information about the collection of the samples (well age, well depth, frequency of use of well, etc.), and analytical data for dysprosium